-- Logs begin at Wed 2018-06-27 14:32:20 EDT. -- Jul 02 16:39:59 pike kubelet[42179]: I0702 16:39:59.700109 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:39:59 pike kubelet[42179]: I0702 16:39:59.940632 42179 eviction_manager.go:229] eviction manager: synchronize housekeeping Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.000539 42179 helpers.go:840] eviction manager: observations: signal=pid.available, available: 129799, capacity: 128Ki, time: 2018-07-02 16:39:59.994099323 -0400 EDT m=+988.527788690 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.000860 42179 helpers.go:840] eviction manager: observations: signal=memory.available, available: 1558296Ki, capacity: 6531384Ki, time: 2018-07-02 16:39:59.958778549 -0400 EDT m=+988.492467950 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.000976 42179 helpers.go:840] eviction manager: observations: signal=allocatableMemory.available, available: 5878156Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:00.000458613 -0400 EDT m=+988.534147975 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.001084 42179 helpers.go:840] eviction manager: observations: signal=nodefs.available, available: 10753296Ki, capacity: 35992192Ki, time: 2018-07-02 16:39:59.958778549 -0400 EDT m=+988.492467950 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.001188 42179 helpers.go:840] eviction manager: observations: signal=nodefs.inodesFree, available: 1688020, capacity: 2240Ki, time: 2018-07-02 16:39:59.958778549 -0400 EDT m=+988.492467950 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.001290 42179 helpers.go:840] eviction manager: observations: signal=imagefs.available, available: 10753296Ki, capacity: 35992192Ki, time: 2018-07-02 16:39:59.958778549 -0400 EDT m=+988.492467950 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.001410 42179 helpers.go:840] eviction manager: observations: signal=imagefs.inodesFree, available: 1688020, capacity: 2240Ki, time: 2018-07-02 16:39:59.958778549 -0400 EDT m=+988.492467950 Jul 02 16:40:00 pike kubelet[42179]: I0702 16:40:00.001592 42179 eviction_manager.go:336] eviction manager: no resources are starved Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.045740 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/kubedns Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.045790 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.046766 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[51] Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:01 GMT]] 0xc420f2cf00 51 [] true false map[] 0xc42058a500 } Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.047034 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.162998 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:etcd-pike,GenerateName:,Namespace:kube-system,SelfLink:/api/v1/namespaces/kube-system/pods/etcd-pike,UID:68f9c58845a7c8f7ac9d6261d498ab6c,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{component: etcd,tier: control-plane,},Annotations:map[string]string{kubernetes.io/config.hash: 68f9c58845a7c8f7ac9d6261d498ab6c,kubernetes.io/config.seen: 2018-07-02T16:23:31.6928551-04:00,kubernetes.io/config.source: file,scheduler.alpha.kubernetes.io/critical-pod: ,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-data {HostPathVolumeSource{Path:/var/lib/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etcd-certs {&HostPathVolumeSource{Path:/etc/kubernetes/pki/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/h Jul 02 16:40:01 pike kubelet[42179]: ealthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:pike,HostNetwork:true,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{ Exists NoExecute }],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds: Jul 02 16:40:01 pike kubelet[42179]: 10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo] Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.238302 42179 exec.go:38] Exec probe response: "2018-07-02 20:40:01.214552 I | warning: ignoring ServerName for user-provided CA for backwards compatibility is deprecated\n" Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.238324 42179 prober.go:118] Liveness probe for "etcd-pike_kube-system(68f9c58845a7c8f7ac9d6261d498ab6c):etcd" succeeded Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.251836 42179 prober.go:165] HTTP-Probe Host: https://192.168.228.17, Port: 6443, Path: /healthz Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.251894 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.257078 42179 http.go:96] Probe succeeded for https://192.168.228.17:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Mon, 02 Jul 2018 20:40:01 GMT]] 0xc4213ca960 2 [] false false map[] 0xc42058af00 0xc4215458c0} Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.257116 42179 prober.go:118] Liveness probe for "kube-apiserver-pike_kube-system(89228f8427ffc79800b43469845b379e):kube-apiserver" succeeded Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.499687 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 8081, Path: /readiness Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.499733 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.500541 42179 http.go:96] Probe succeeded for http://192.168.0.43:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:01 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc4213caa60 3 [] true false map[] 0xc42058b200 } Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.500581 42179 prober.go:118] Readiness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.524543 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.524603 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.525641 42179 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Mon, 02 Jul 2018 20:40:01 GMT] Content-Length:[2]] 0xc421ccab80 2 [] true false map[] 0xc420011500 } Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.525689 42179 prober.go:118] Liveness probe for "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e):kube-scheduler" succeeded Jul 02 16:40:01 pike kubelet[42179]: I0702 16:40:01.700393 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219009 42179 interface.go:360] Looking for default routes with IPv4 addresses Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219048 42179 interface.go:365] Default route transits interface "ens33" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219720 42179 interface.go:174] Interface ens33 is up Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219776 42179 interface.go:222] Interface "ens33" has 2 addresses :[192.168.228.17/24 fe80::779f:8935:8254:8bd8/64]. Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219793 42179 interface.go:189] Checking addr 192.168.228.17/24. Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219800 42179 interface.go:196] IP found 192.168.228.17 Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219805 42179 interface.go:228] Found valid IPv4 address 192.168.228.17 for interface "ens33". Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.219810 42179 interface.go:371] Found active IP 192.168.228.17 Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.388957 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /readiness Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.389422 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.391741 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/readiness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:03 GMT]] 0xc4214c80c0 0 [] true false map[] 0xc420c87500 } Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.391829 42179 prober.go:118] Readiness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.620839 42179 config.go:297] Setting pods for source api Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.621360 42179 config.go:405] Receiving a new pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.621430 42179 kubelet.go:1856] SyncLoop (ADD, "api"): "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.622828 42179 kubelet_pods.go:1366] Generating status for "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.623282 42179 qos_container_manager_linux.go:317] [ContainerManager]: Updated QoS cgroup configuration Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.625844 42179 factory.go:117] Factory "docker" was unable to handle container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.625864 42179 factory.go:106] Error trying to work out if we can handle /kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598: /kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598 not handled by systemd handler Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.625869 42179 factory.go:117] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.625874 42179 factory.go:113] Using factory "raw" for container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.625972 42179 manager.go:997] Added container: "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" (aliases: [], namespace: "") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.626024 42179 handler.go:325] Added event &{/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598 2018-07-02 16:40:03.623101187 -0400 EDT containerCreation {}} Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.626051 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.631932 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.633971 42179 config.go:297] Setting pods for source api Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.634898 42179 kubelet.go:1869] SyncLoop (RECONCILE, "api"): "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.635730 42179 status_manager.go:481] Status for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" updated successfully: (1, {Phase:Pending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP: StartTime:2018-07-02 16:40:03 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:nginx:latest ImageID: ContainerID:}] QOSClass:BestEffort}) Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.699876 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.711082 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "18ea1a7c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.711153 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "18ea1a7c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.731024 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.731077 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:03 pike kubelet[42179]: E0702 16:40:03.731169 42179 nestedpendingoperations.go:267] Operation for "\"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\"" failed. No retries permitted until 2018-07-02 16:40:04.231145744 -0400 EDT m=+992.764835089 (durationBeforeRetry 500ms). Error: "Volume has not been added to the list of VolumesInUse in the node's volume status for volume \"pvc-281709ce7e3711e8\" (UniqueName: \"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\") pod \"pod2\" (UID: \"18ea1a7c-7e38-11e8-a286-005056383598\") " Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.831786 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.831883 42179 reconciler.go:252] operationExecutor.MountVolume started for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.831999 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 18ea1a7c-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.832210 42179 empty_dir.go:256] pod 18ea1a7c-7e38-11e8-a286-005056383598: mounting tmpfs for volume wrapped_default-token-9dv6w Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.832239 42179 mount_linux.go:143] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w]) Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.837643 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.837687 42179 factory.go:106] Error trying to work out if we can handle /system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope: /system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope not handled by systemd handler Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.837694 42179 factory.go:117] Factory "systemd" was unable to handle container "/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.837699 42179 factory.go:113] Using factory "raw" for container "/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.837900 42179 manager.go:997] Added container: "/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope" (aliases: [], namespace: "") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.837992 42179 handler.go:325] Added event &{/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope 2018-07-02 16:40:03.835091605 -0400 EDT containerCreation {}} Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.838029 42179 container.go:448] Start housekeeping for container "/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope" Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.848546 42179 manager.go:1054] Destroyed container: "/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope" (aliases: [], namespace: "") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.848940 42179 handler.go:325] Added event &{/system.slice/run-r8b74632156754e6ba29f87b9565576bb.scope 2018-07-02 16:40:03.848930244 -0400 EDT m=+992.382619610 containerDeletion {}} Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.850979 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.851082 42179 atomic_writer.go:176] pod trident/pod2 volume default-token-9dv6w: performed write of new data to ts data directory: /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w/..2018_07_02_20_40_03.910949543 Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.851181 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:03 pike kubelet[42179]: I0702 16:40:03.851647 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-9dv6w" Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.144694 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /liveness Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.144739 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.145800 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/liveness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:04 GMT]] 0xc421a415a0 0 [] true false map[] 0xc420a9ad00 } Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.145856 42179 prober.go:118] Liveness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.234440 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:04 pike kubelet[42179]: E0702 16:40:04.234560 42179 nestedpendingoperations.go:267] Operation for "\"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\"" failed. No retries permitted until 2018-07-02 16:40:05.234507318 -0400 EDT m=+993.768196667 (durationBeforeRetry 1s). Error: "Volume has not been added to the list of VolumesInUse in the node's volume status for volume \"pvc-281709ce7e3711e8\" (UniqueName: \"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\") pod \"pod2\" (UID: \"18ea1a7c-7e38-11e8-a286-005056383598\") " Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.614319 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /metrics Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.614455 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.621354 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; version=0.0.4] Date:[Mon, 02 Jul 2018 20:40:04 GMT]] 0xc421153be0 -1 [] true true map[] 0xc420a9b000 } Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.621425 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):sidecar" succeeded Jul 02 16:40:04 pike kubelet[42179]: I0702 16:40:04.666260 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.242195 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:05 pike kubelet[42179]: E0702 16:40:05.242339 42179 nestedpendingoperations.go:267] Operation for "\"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\"" failed. No retries permitted until 2018-07-02 16:40:07.242299893 -0400 EDT m=+995.775989248 (durationBeforeRetry 2s). Error: "Volume has not been added to the list of VolumesInUse in the node's volume status for volume \"pvc-281709ce7e3711e8\" (UniqueName: \"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\") pod \"pod2\" (UID: \"18ea1a7c-7e38-11e8-a286-005056383598\") " Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.701856 42179 kubelet.go:1920] SyncLoop (SYNC): 1 pods; calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598) Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.701993 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.702570 42179 kubelet_pods.go:1366] Generating status for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598)" Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.702846 42179 status_manager.go:340] Ignoring same status for pod "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-23 19:13:08 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:23:43 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-23 19:13:08 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.228.17 StartTime:2018-06-23 19:13:08 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:calico-node State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-06-29 12:55:58 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2018-06-29 11:42:55 -0400 EDT,FinishedAt:2018-06-29 12:29:24 -0400 EDT,ContainerID:docker://70577d70edbd78f1cb0c291c5d79e3c8f42918e3580aa6595dc41143419e48a9,}} Ready:true RestartCount:2 Image:quay.io/calico/node:v3.1.3 ImageID:docker-pullable://quay.io/calico/node@sha256:a35541153f7695b38afada46843c64a2c546548cd8c171f402621736c6cf3f0b ContainerID:docker://7f107bdfe0ff2ffec79140650498bf47d6730e4531a7da560dcaf0217781a9fe} {Name:install-cni State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-06-29 12:56:15 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2018-06-29 11:43:11 -0400 EDT,FinishedAt:2018-06-29 12:50:30 -0400 EDT,ContainerID:docker://8d9c83b933cae7d9b072549ecaabde0586c5d18e0ea2ab1273e486344bfc514c,}} Ready:true RestartCount:2 Image:quay.io/calico/cni:v3.1.3 ImageID:docker-pullable://quay.io/calico/cni@sha256:ed172c28bc193bb09bce6be6ed7dc6bfc85118d55e61d263cee8bb Jul 02 16:40:05 pike kubelet[42179]: b0fd464a9d ContainerID:docker://e3ad4a5d486f8415fd9a9671e37ef795f04624b62a18f9c82b1b3cccfe995720}] QOSClass:Burstable} Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.703166 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598)" Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.726065 42179 desired_state_of_world_populator.go:302] Added volume "lib-modules" (volSpec="lib-modules") for pod "fe1240e4-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.726125 42179 desired_state_of_world_populator.go:302] Added volume "var-run-calico" (volSpec="var-run-calico") for pod "fe1240e4-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.726150 42179 desired_state_of_world_populator.go:302] Added volume "var-lib-calico" (volSpec="var-lib-calico") for pod "fe1240e4-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.726160 42179 desired_state_of_world_populator.go:302] Added volume "cni-bin-dir" (volSpec="cni-bin-dir") for pod "fe1240e4-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.726171 42179 desired_state_of_world_populator.go:302] Added volume "cni-net-dir" (volSpec="cni-net-dir") for pod "fe1240e4-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.726185 42179 desired_state_of_world_populator.go:302] Added volume "calico-node-token-vrxwt" (volSpec="calico-node-token-vrxwt") for pod "fe1240e4-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.747535 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "calico-node-token-vrxwt" (UniqueName: "kubernetes.io/secret/fe1240e4-773a-11e8-a95f-005056383598-calico-node-token-vrxwt") pod "calico-node-fnnvb" (UID: "fe1240e4-773a-11e8-a95f-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.747739 42179 secret.go:186] Setting up volume calico-node-token-vrxwt for pod fe1240e4-773a-11e8-a95f-005056383598 at /var/lib/kubelet/pods/fe1240e4-773a-11e8-a95f-005056383598/volumes/kubernetes.io~secret/calico-node-token-vrxwt Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.749970 42179 secret.go:216] Received secret kube-system/calico-node-token-vrxwt containing (3) pieces of data, 1904 total bytes Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.750168 42179 atomic_writer.go:156] pod kube-system/calico-node-fnnvb volume calico-node-token-vrxwt: no update required for target directory /var/lib/kubelet/pods/fe1240e4-773a-11e8-a95f-005056383598/volumes/kubernetes.io~secret/calico-node-token-vrxwt Jul 02 16:40:05 pike kubelet[42179]: I0702 16:40:05.750203 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "calico-node-token-vrxwt" (UniqueName: "kubernetes.io/secret/fe1240e4-773a-11e8-a95f-005056383598-calico-node-token-vrxwt") pod "calico-node-fnnvb" (UID: "fe1240e4-773a-11e8-a95f-005056383598") Jul 02 16:40:06 pike kubelet[42179]: I0702 16:40:06.003889 42179 volume_manager.go:380] All volumes are attached and mounted for pod "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598)" Jul 02 16:40:06 pike kubelet[42179]: I0702 16:40:06.004176 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d716714af9ce066cc00c61812e99a57bb6321a94cd5ea4e5cd804a496d42fef2 Attempt:2 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598)" Jul 02 16:40:07 pike kubelet[42179]: I0702 16:40:07.030205 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz Jul 02 16:40:07 pike kubelet[42179]: I0702 16:40:07.030325 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:07 pike kubelet[42179]: I0702 16:40:07.031956 42179 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Content-Type:[text/plain; charset=utf-8] Date:[Mon, 02 Jul 2018 20:40:07 GMT]] 0xc422abe7a0 2 [] true false map[] 0xc4215e2400 } Jul 02 16:40:07 pike kubelet[42179]: I0702 16:40:07.032066 42179 prober.go:118] Liveness probe for "kube-controller-manager-pike_kube-system(fac56382eca21a00fc5c8a039563b311):kube-controller-manager" succeeded Jul 02 16:40:07 pike kubelet[42179]: I0702 16:40:07.257944 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:07 pike kubelet[42179]: E0702 16:40:07.258071 42179 nestedpendingoperations.go:267] Operation for "\"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\"" failed. No retries permitted until 2018-07-02 16:40:11.25803344 -0400 EDT m=+999.791722804 (durationBeforeRetry 4s). Error: "Volume has not been added to the list of VolumesInUse in the node's volume status for volume \"pvc-281709ce7e3711e8\" (UniqueName: \"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\") pod \"pod2\" (UID: \"18ea1a7c-7e38-11e8-a286-005056383598\") " Jul 02 16:40:07 pike kubelet[42179]: I0702 16:40:07.700102 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:09 pike kubelet[42179]: I0702 16:40:09.207279 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/dnsmasq Jul 02 16:40:09 pike kubelet[42179]: I0702 16:40:09.207411 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:09 pike kubelet[42179]: I0702 16:40:09.208950 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[50] Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:09 GMT]] 0xc420741400 50 [] true false map[] 0xc422b62400 } Jul 02 16:40:09 pike kubelet[42179]: I0702 16:40:09.209053 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):dnsmasq" succeeded Jul 02 16:40:09 pike kubelet[42179]: I0702 16:40:09.668209 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:09 pike kubelet[42179]: I0702 16:40:09.699923 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.002322 42179 eviction_manager.go:229] eviction manager: synchronize housekeeping Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056096 42179 helpers.go:840] eviction manager: observations: signal=nodefs.inodesFree, available: 1688011, capacity: 2240Ki, time: 2018-07-02 16:40:10.019666674 -0400 EDT m=+998.553356060 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056149 42179 helpers.go:840] eviction manager: observations: signal=imagefs.available, available: 10753168Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:10.019666674 -0400 EDT m=+998.553356060 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056160 42179 helpers.go:840] eviction manager: observations: signal=imagefs.inodesFree, available: 1688011, capacity: 2240Ki, time: 2018-07-02 16:40:10.019666674 -0400 EDT m=+998.553356060 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056166 42179 helpers.go:840] eviction manager: observations: signal=pid.available, available: 129798, capacity: 128Ki, time: 2018-07-02 16:40:10.051952111 -0400 EDT m=+998.585641458 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056171 42179 helpers.go:840] eviction manager: observations: signal=memory.available, available: 1556816Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:10.019666674 -0400 EDT m=+998.553356060 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056177 42179 helpers.go:840] eviction manager: observations: signal=allocatableMemory.available, available: 5877272Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:10.055971399 -0400 EDT m=+998.589660764 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056182 42179 helpers.go:840] eviction manager: observations: signal=nodefs.available, available: 10753168Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:10.019666674 -0400 EDT m=+998.553356060 Jul 02 16:40:10 pike kubelet[42179]: I0702 16:40:10.056230 42179 eviction_manager.go:336] eviction manager: no resources are starved Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.044524 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/kubedns Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.045214 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.047048 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:11 GMT] Content-Length:[51]] 0xc4228e8ac0 51 [] true false map[] 0xc42185c600 } Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.047319 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.163276 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:etcd-pike,GenerateName:,Namespace:kube-system,SelfLink:/api/v1/namespaces/kube-system/pods/etcd-pike,UID:68f9c58845a7c8f7ac9d6261d498ab6c,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{component: etcd,tier: control-plane,},Annotations:map[string]string{kubernetes.io/config.hash: 68f9c58845a7c8f7ac9d6261d498ab6c,kubernetes.io/config.seen: 2018-07-02T16:23:31.6928551-04:00,kubernetes.io/config.source: file,scheduler.alpha.kubernetes.io/critical-pod: ,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-data {HostPathVolumeSource{Path:/var/lib/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etcd-certs {&HostPathVolumeSource{Path:/etc/kubernetes/pki/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/h Jul 02 16:40:11 pike kubelet[42179]: ealthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:pike,HostNetwork:true,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{ Exists NoExecute }],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds: Jul 02 16:40:11 pike kubelet[42179]: 10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo] Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.239516 42179 exec.go:38] Exec probe response: "2018-07-02 20:40:11.213473 I | warning: ignoring ServerName for user-provided CA for backwards compatibility is deprecated\n" Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.239845 42179 prober.go:118] Liveness probe for "etcd-pike_kube-system(68f9c58845a7c8f7ac9d6261d498ab6c):etcd" succeeded Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.252592 42179 prober.go:165] HTTP-Probe Host: https://192.168.228.17, Port: 6443, Path: /healthz Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.252959 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.258508 42179 http.go:96] Probe succeeded for https://192.168.228.17:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Mon, 02 Jul 2018 20:40:11 GMT]] 0xc4213cbb00 2 [] false false map[] 0xc42185d000 0xc421382840} Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.258550 42179 prober.go:118] Liveness probe for "kube-apiserver-pike_kube-system(89228f8427ffc79800b43469845b379e):kube-apiserver" succeeded Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.286911 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:11 pike kubelet[42179]: E0702 16:40:11.287426 42179 nestedpendingoperations.go:267] Operation for "\"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\"" failed. No retries permitted until 2018-07-02 16:40:19.28732174 -0400 EDT m=+1007.821011100 (durationBeforeRetry 8s). Error: "Volume has not been added to the list of VolumesInUse in the node's volume status for volume \"pvc-281709ce7e3711e8\" (UniqueName: \"kubernetes.io/csi/io.netapp.trident.csi^{\\\"name\\\":\\\"pvc-281709ce7e3711e8\\\",\\\"protocol\\\":\\\"block\\\"}\") pod \"pod2\" (UID: \"18ea1a7c-7e38-11e8-a286-005056383598\") " Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.499627 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 8081, Path: /readiness Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.499627 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 8081, Path: /readiness Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.500317 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.501188 42179 http.go:96] Probe succeeded for http://192.168.0.43:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:11 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc422a82cc0 3 [] true false map[] 0xc4226b0800 } Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.501221 42179 prober.go:118] Readiness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.525130 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.525228 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.526500 42179 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:11 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421feed20 2 [] true false map[] 0xc422b63900 } Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.526591 42179 prober.go:118] Liveness probe for "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e):kube-scheduler" succeeded Jul 02 16:40:11 pike kubelet[42179]: I0702 16:40:11.700099 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:12 pike kubelet[42179]: I0702 16:40:12.213447 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:trident-csi-0,GenerateName:trident-csi-,Namespace:trident,SelfLink:/api/v1/namespaces/trident/pods/trident-csi-0,UID:8eaa19b4-7e34-11e8-a286-005056383598,ResourceVersion:147708,Generation:0,CreationTimestamp:2018-07-02 16:14:43 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{app: controller.csi.trident.netapp.io,controller-revision-hash: trident-csi-66f8f9d5bb,statefulset.kubernetes.io/pod-name: trident-csi-0,},Annotations:map[string]string{cni.projectcalico.org/podIP: 192.168.0.49/32,kubernetes.io/config.seen: 2018-07-02T16:23:31.650583191-04:00,kubernetes.io/config.source: api,},OwnerReferences:[{apps/v1 StatefulSet trident-csi 8e98152e-7e34-11e8-a286-005056383598 0xc420438d08 0xc420438d09}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-vol {nil nil nil nil nil nil nil nil nil PersistentVolumeClaimVolumeSource{ClaimName:trident-csi,ReadOnly:false,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {socket-dir {nil &EmptyDirVolumeSource{Medium:,SizeLimit:,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etc-dir {&HostPathVolumeSource{Path:/etc,Type:*Directory,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {trident-csi-token-4gzh4 {nil nil nil nil nil &SecretVolumeSource{SecretName:trident-csi-token-4gzh4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{trident-main registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae [/usr/local/bin/trident_orchestrator] [-etcd_v3 http://127.0.0.1:8001 --csi_node_name=$(KUBE_NODE_NAME) --csi_endpoint=$(CSI_ENDPOINT) -debug] [] [] [{KUBE_NODE_NAME EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldP Jul 02 16:40:12 pike kubelet[42179]: ath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,}} {CSI_ENDPOINT unix://plugin/csi.sock nil}] {map[] map[]} [{socket-dir false /plugin } {etc-dir false /etc } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[tridentctl -s 127.0.0.1:8000 get backend],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:120,TimeoutSeconds:90,PeriodSeconds:120,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false} {etcd quay.io/coreos/etcd:v3.2.19 [/usr/local/bin/etcd] [-name etcd1 -advertise-client-urls http://127.0.0.1:8001 -listen-client-urls http://127.0.0.1:8001 -initial-advertise-peer-urls http://127.0.0.1:8002 -listen-peer-urls http://127.0.0.1:8002 -data-dir /var/etcd/data -initial-cluster etcd1=http://127.0.0.1:8002] [] [] [] {map[] map[]} [{etcd-vol false /var/etcd/data } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:10,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false} {csi-attacher quay.io/k8scsi/csi-attacher:v0.2.0 [] [--v=9 --csi-address=$(ADDRESS)] [] [] [{ADDRESS /var/lib/csi/sockets/pluginproxy/csi.sock nil}] {map[] map[]} [{socket-dir false /var/lib/csi/sockets/pluginproxy/ } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false} {csi-provisioner quay.io/k8scsi/csi-provisioner:v0.2.1 [] [--v=9 --provisioner=io.netapp.trident.csi --csi-address=$(ADDRESS)] [] [] [{ADDRESS /var/lib/csi/sockets/pluginproxy/csi.sock nil}] {map[] map[]} [{socket-dir false /var/lib/csi/sockets/pluginproxy/ } {trident-csi-token-4gzh4 true /var/run/s Jul 02 16:40:12 pike kubelet[42179]: ecrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:trident-csi,DeprecatedServiceAccount:trident-csi,NodeName:pike,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:trident-csi-0,Subdomain:trident-csi,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc420438d18} {node.kubernetes.io/unreachable Exists NoExecute 0xc420438d20}],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:43 -0400 EDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:48 -0400 EDT } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:43 -0400 EDT }],Message:,Reason:,HostIP:192.168.228.17,PodIP:192.168.0.49,StartTime:2018-07-02 16:14:43 -0400 EDT,ContainerStatuses:[{csi-attacher {nil ContainerStateRunning{StartedAt:2018-07-02 16:14:47 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/k8scsi/csi-attacher:v0.2.0 docker-pullable://quay.io/k8scsi/csi-attacher@sha256:5cbb7934bd86d400c221379cff8b24ed4c06e121ea59608cfd7e67690100ba54 docker://f547da90c65d01207a828ee3b8d044eb4b4cc05fa065f39648eac3e539f3b9a6} {csi-provisioner {nil &ContainerStateRunning{StartedAt:2018-07-02 16:14:48 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/k8scsi/csi-provisioner:v0.2.1 docker-pullable://quay.io/k8scsi/csi-provisioner@sha256:fd4ed32315e846b6654f97c95b373da001fd9638cd5935c20a5bf9f5889e8602 docker://08d31d6c565ec486d8e9c46017c10edd0f390da6a139692a9daa25e620777bc0} {etcd {nil &ContainerStateRunning{StartedAt:2018-0 Jul 02 16:40:12 pike kubelet[42179]: 7-02 16:14:47 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/coreos/etcd:v3.2.19 docker-pullable://quay.io/coreos/etcd@sha256:e21b9734136122b1e8b9b37742e72d2bd8b0d84bca875892494e0de5520313b0 docker://8340342e4a7288ef0672e62be7c0bf34e1800f7ce961607fa163146aa828b861} {trident-main {nil &ContainerStateRunning{StartedAt:2018-07-02 16:14:46 -0400 EDT,} nil} {nil nil nil} true 0 registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae docker-pullable://registry.barnacle.netapp.com:5000/trident@sha256:ce8cd402f9ef85e18ce428f09137a7347c5009c255daf730f9c40c006ea0ee23 docker://2c51e41d415063f9be931cf9ebbb62bf4d4c427d9002f312fbf54f9c34a4d9ee}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd quay.io/coreos/etcd:v3.2.19 [/usr/local/bin/etcd] [-name etcd1 -advertise-client-urls http://127.0.0.1:8001 -listen-client-urls http://127.0.0.1:8001 -initial-advertise-peer-urls http://127.0.0.1:8002 -listen-peer-urls http://127.0.0.1:8002 -data-dir /var/etcd/data -initial-cluster etcd1=http://127.0.0.1:8002] [] [] [] {map[] map[]} [{etcd-vol false /var/etcd/data } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:10,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health] Jul 02 16:40:12 pike kubelet[42179]: I0702 16:40:12.650257 42179 exec.go:38] Exec probe response: "member 9ca0c7938a0c3dc5 is healthy: got healthy result from http://127.0.0.1:8001\ncluster is healthy\n" Jul 02 16:40:12 pike kubelet[42179]: I0702 16:40:12.650309 42179 prober.go:118] Liveness probe for "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598):etcd" succeeded Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231141 42179 interface.go:360] Looking for default routes with IPv4 addresses Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231162 42179 interface.go:365] Default route transits interface "ens33" Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231780 42179 interface.go:174] Interface ens33 is up Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231906 42179 interface.go:222] Interface "ens33" has 2 addresses :[192.168.228.17/24 fe80::779f:8935:8254:8bd8/64]. Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231927 42179 interface.go:189] Checking addr 192.168.228.17/24. Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231939 42179 interface.go:196] IP found 192.168.228.17 Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231949 42179 interface.go:228] Found valid IPv4 address 192.168.228.17 for interface "ens33". Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.231983 42179 interface.go:371] Found active IP 192.168.228.17 Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.388529 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /readiness Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.388562 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.389383 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/readiness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:13 GMT]] 0xc421ccaf80 0 [] true false map[] 0xc42058ab00 } Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.389591 42179 prober.go:118] Readiness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:13 pike kubelet[42179]: I0702 16:40:13.700016 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.145254 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /liveness Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.145305 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.147069 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/liveness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:14 GMT]] 0xc420c67da0 0 [] true false map[] 0xc421203200 } Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.147142 42179 prober.go:118] Liveness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.614303 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /metrics Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.614390 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.622194 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; version=0.0.4] Date:[Mon, 02 Jul 2018 20:40:14 GMT]] 0xc421244ca0 -1 [] true true map[] 0xc420a9a900 } Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.622442 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):sidecar" succeeded Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.671780 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.700268 42179 kubelet.go:1920] SyncLoop (SYNC): 1 pods; trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598) Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.700403 42179 kubelet_pods.go:1366] Generating status for "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598)" Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.700604 42179 status_manager.go:340] Ignoring same status for pod "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:14:43 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:14:48 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:14:43 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.49 StartTime:2018-07-02 16:14:43 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:csi-attacher State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:14:47 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:quay.io/k8scsi/csi-attacher:v0.2.0 ImageID:docker-pullable://quay.io/k8scsi/csi-attacher@sha256:5cbb7934bd86d400c221379cff8b24ed4c06e121ea59608cfd7e67690100ba54 ContainerID:docker://f547da90c65d01207a828ee3b8d044eb4b4cc05fa065f39648eac3e539f3b9a6} {Name:csi-provisioner State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:14:48 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:quay.io/k8scsi/csi-provisioner:v0.2.1 ImageID:docker-pullable://quay.io/k8scsi/csi-provisioner@sha256:fd4ed32315e846b6654f97c95b373da001fd9638cd5935c20a5bf9f5889e8602 ContainerID:docker://08d31d6c565ec486d8e9c46017c10edd0f390da6a139692a9daa25e620777bc0} {Name:etcd State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:14:47 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:quay.io/coreos/etcd:v3.2.19 ImageID:docker-pullable://quay.io/coreos/etcd@sha256:e21b9734136122b1e8b9b37742e72d2bd8b0d84bca87589 Jul 02 16:40:14 pike kubelet[42179]: 2494e0de5520313b0 ContainerID:docker://8340342e4a7288ef0672e62be7c0bf34e1800f7ce961607fa163146aa828b861} {Name:trident-main State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:14:46 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae ImageID:docker-pullable://registry.barnacle.netapp.com:5000/trident@sha256:ce8cd402f9ef85e18ce428f09137a7347c5009c255daf730f9c40c006ea0ee23 ContainerID:docker://2c51e41d415063f9be931cf9ebbb62bf4d4c427d9002f312fbf54f9c34a4d9ee}] QOSClass:BestEffort} Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.700828 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598)" Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.789894 42179 desired_state_of_world_populator.go:302] Added volume "etcd-vol" (volSpec="trident-csi") for pod "8eaa19b4-7e34-11e8-a286-005056383598" to desired state. Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.789935 42179 desired_state_of_world_populator.go:302] Added volume "socket-dir" (volSpec="socket-dir") for pod "8eaa19b4-7e34-11e8-a286-005056383598" to desired state. Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.789950 42179 desired_state_of_world_populator.go:302] Added volume "etc-dir" (volSpec="etc-dir") for pod "8eaa19b4-7e34-11e8-a286-005056383598" to desired state. Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.789965 42179 desired_state_of_world_populator.go:302] Added volume "trident-csi-token-4gzh4" (volSpec="trident-csi-token-4gzh4") for pod "8eaa19b4-7e34-11e8-a286-005056383598" to desired state. Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.810112 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "trident-csi-token-4gzh4" (UniqueName: "kubernetes.io/secret/8eaa19b4-7e34-11e8-a286-005056383598-trident-csi-token-4gzh4") pod "trident-csi-0" (UID: "8eaa19b4-7e34-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.810284 42179 secret.go:186] Setting up volume trident-csi-token-4gzh4 for pod 8eaa19b4-7e34-11e8-a286-005056383598 at /var/lib/kubelet/pods/8eaa19b4-7e34-11e8-a286-005056383598/volumes/kubernetes.io~secret/trident-csi-token-4gzh4 Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.811958 42179 secret.go:216] Received secret trident/trident-csi-token-4gzh4 containing (3) pieces of data, 1890 total bytes Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.812395 42179 atomic_writer.go:156] pod trident/trident-csi-0 volume trident-csi-token-4gzh4: no update required for target directory /var/lib/kubelet/pods/8eaa19b4-7e34-11e8-a286-005056383598/volumes/kubernetes.io~secret/trident-csi-token-4gzh4 Jul 02 16:40:14 pike kubelet[42179]: I0702 16:40:14.812663 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "trident-csi-token-4gzh4" (UniqueName: "kubernetes.io/secret/8eaa19b4-7e34-11e8-a286-005056383598-trident-csi-token-4gzh4") pod "trident-csi-0" (UID: "8eaa19b4-7e34-11e8-a286-005056383598") Jul 02 16:40:15 pike kubelet[42179]: I0702 16:40:15.001249 42179 volume_manager.go:380] All volumes are attached and mounted for pod "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598)" Jul 02 16:40:15 pike kubelet[42179]: I0702 16:40:15.001421 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fdf93005011de0e9a6e99298bacb4068da25a044851e342c364e9a1807acc085 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598)" Jul 02 16:40:15 pike kubelet[42179]: I0702 16:40:15.700321 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:17 pike kubelet[42179]: I0702 16:40:17.029984 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz Jul 02 16:40:17 pike kubelet[42179]: I0702 16:40:17.030084 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:17 pike kubelet[42179]: I0702 16:40:17.031771 42179 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Content-Type:[text/plain; charset=utf-8] Date:[Mon, 02 Jul 2018 20:40:17 GMT]] 0xc420d4b8e0 2 [] true false map[] 0xc421203f00 } Jul 02 16:40:17 pike kubelet[42179]: I0702 16:40:17.031841 42179 prober.go:118] Liveness probe for "kube-controller-manager-pike_kube-system(fac56382eca21a00fc5c8a039563b311):kube-controller-manager" succeeded Jul 02 16:40:17 pike kubelet[42179]: I0702 16:40:17.700302 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.207627 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/dnsmasq Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.207683 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.209080 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:19 GMT] Content-Length:[51]] 0xc422b43460 51 [] true false map[] 0xc420011500 } Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.209134 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):dnsmasq" succeeded Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.347625 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.351106 42179 operation_generator.go:1168] Controller attach succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") device path: "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.448292 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.448801 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.449101 42179 reconciler.go:252] operationExecutor.MountVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.449449 42179 operation_generator.go:486] MountVolume.WaitForAttach entering for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") DevicePath "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.449537 42179 csi_attacher.go:131] kubernetes.io/csi: probing for updates from CSI driver for [attachment.ID=csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.449576 42179 csi_attacher.go:140] kubernetes.io/csi: probing VolumeAttachment [id=csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.453917 42179 operation_generator.go:495] MountVolume.WaitForAttach succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") DevicePath "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.454065 42179 csi_attacher.go:244] kubernetes.io/csi: attacher.GetDeviceMountPath(&{nil &PersistentVolume{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:pvc-281709ce7e3711e8,GenerateName:,Namespace:,SelfLink:/api/v1/persistentvolumes/pvc-281709ce7e3711e8,UID:2c19eed1-7e37-11e8-a286-005056383598,ResourceVersion:149392,Generation:0,CreationTimestamp:2018-07-02 16:33:26 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{pv.kubernetes.io/provisioned-by: io.netapp.trident.csi,},OwnerReferences:[],Finalizers:[kubernetes.io/pv-protection external-attacher/io-netapp-trident-csi],ClusterName:,Initializers:nil,},Spec:PersistentVolumeSpec{Capacity:ResourceList{storage: {{1073741824 0} {} 1Gi BinarySI},},PersistentVolumeSource:PersistentVolumeSource{GCEPersistentDisk:nil,AWSElasticBlockStore:nil,HostPath:nil,Glusterfs:nil,NFS:nil,RBD:nil,ISCSI:nil,Cinder:nil,CephFS:nil,FC:nil,Flocker:nil,FlexVolume:nil,AzureFile:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Local:nil,StorageOS:nil,CSI:&CSIPersistentVolumeSource{Driver:io.netapp.trident.csi,VolumeHandle:{"name":"pvc-281709ce7e3711e8","protocol":"block"},ReadOnly:false,FSType:,VolumeAttributes:map[string]string{backend: solidfire_10.250.118.144,internalName: pvc-281709ce7e3711e8,name: pvc-281709ce7e3711e8,protocol: block,storage.kubernetes.io/csiProvisionerIdentity: 1530562488326-8081-io.netapp.trident.csi,},ControllerPublishSecretRef:nil,NodeStageSecretRef:nil,NodePublishSecretRef:nil,},},AccessModes:[ReadWriteMany],ClaimRef:&ObjectReference{Kind:PersistentVolumeClaim,Namespace:trident,Name:pvc2,UID:281709ce-7e37-11e8-a286-005056383598,APIVersion:v1,ResourceVersion:149109,FieldPath:,},PersistentVolumeReclaimPolicy:Delete,StorageClassName:ssd-sf,MountOptions:[],VolumeMode:nil,NodeAffinity:nil,},Status:PersistentVolumeStatus{Phase:Bound,Message:,Reason:,},} false}) Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.454085 42179 csi_attacher.go:250] attacher.GetDeviceMountPath succeeded, deviceMountPath: /var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.454093 42179 csi_attacher.go:255] kubernetes.io/csi: attacher.MountDevice(csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c, /var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount) Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.454123 42179 csi_client.go:242] kubernetes.io/csi: calling NodeGetCapabilities rpc Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.457419 42179 csi_attacher.go:326] kubernetes.io/csi: created target path successfully [/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.457464 42179 csi_client.go:186] kubernetes.io/csi: calling NodeStageVolume rpc [volid={"name":"pvc-281709ce7e3711e8","protocol":"block"},staging_target_path=/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.550147 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.550291 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.651001 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.651129 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.674949 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.700160 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.753254 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.753313 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.854347 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.854478 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.955597 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:19 pike kubelet[42179]: I0702 16:40:19.955756 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.056405 42179 eviction_manager.go:229] eviction manager: synchronize housekeeping Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.056496 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.056572 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.113428 42179 helpers.go:840] eviction manager: observations: signal=memory.available, available: 1552216Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:20.076111292 -0400 EDT m=+1008.609800663 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.113885 42179 helpers.go:840] eviction manager: observations: signal=allocatableMemory.available, available: 5874884Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:20.11331618 -0400 EDT m=+1008.647005541 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.114056 42179 helpers.go:840] eviction manager: observations: signal=nodefs.available, available: 10753056Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:20.076111292 -0400 EDT m=+1008.609800663 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.114229 42179 helpers.go:840] eviction manager: observations: signal=nodefs.inodesFree, available: 1688009, capacity: 2240Ki, time: 2018-07-02 16:40:20.076111292 -0400 EDT m=+1008.609800663 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.114397 42179 helpers.go:840] eviction manager: observations: signal=imagefs.available, available: 10753056Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:20.076111292 -0400 EDT m=+1008.609800663 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.114562 42179 helpers.go:840] eviction manager: observations: signal=imagefs.inodesFree, available: 1688009, capacity: 2240Ki, time: 2018-07-02 16:40:20.076111292 -0400 EDT m=+1008.609800663 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.114731 42179 helpers.go:840] eviction manager: observations: signal=pid.available, available: 129788, capacity: 128Ki, time: 2018-07-02 16:40:20.10948873 -0400 EDT m=+1008.643178083 Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.114954 42179 eviction_manager.go:336] eviction manager: no resources are starved Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.157447 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.157990 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.258798 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.258868 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.359655 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.359984 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.460710 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.460825 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.561409 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.561813 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.662592 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.662875 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.764170 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.764356 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.866582 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.867011 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.967992 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:20 pike kubelet[42179]: I0702 16:40:20.968532 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.044828 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/kubedns Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.044915 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.045870 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:21 GMT] Content-Length:[51] Content-Type:[application/json]] 0xc4214c8400 51 [] true false map[] 0xc4203b5c00 } Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.045943 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.069462 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.069569 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.164058 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:etcd-pike,GenerateName:,Namespace:kube-system,SelfLink:/api/v1/namespaces/kube-system/pods/etcd-pike,UID:68f9c58845a7c8f7ac9d6261d498ab6c,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{component: etcd,tier: control-plane,},Annotations:map[string]string{kubernetes.io/config.hash: 68f9c58845a7c8f7ac9d6261d498ab6c,kubernetes.io/config.seen: 2018-07-02T16:23:31.6928551-04:00,kubernetes.io/config.source: file,scheduler.alpha.kubernetes.io/critical-pod: ,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-data {HostPathVolumeSource{Path:/var/lib/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etcd-certs {&HostPathVolumeSource{Path:/etc/kubernetes/pki/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/h Jul 02 16:40:21 pike kubelet[42179]: ealthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:pike,HostNetwork:true,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{ Exists NoExecute }],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds: Jul 02 16:40:21 pike kubelet[42179]: 10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.169935 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.170192 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.243480 42179 exec.go:38] Exec probe response: "2018-07-02 20:40:21.220438 I | warning: ignoring ServerName for user-provided CA for backwards compatibility is deprecated\n" Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.243719 42179 prober.go:118] Liveness probe for "etcd-pike_kube-system(68f9c58845a7c8f7ac9d6261d498ab6c):etcd" succeeded Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.251943 42179 prober.go:165] HTTP-Probe Host: https://192.168.228.17, Port: 6443, Path: /healthz Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.252034 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.258914 42179 http.go:96] Probe succeeded for https://192.168.228.17:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Date:[Mon, 02 Jul 2018 20:40:21 GMT] Content-Type:[text/plain; charset=utf-8] Content-Length:[2]] 0xc421b84320 2 [] false false map[] 0xc4228c2100 0xc421f6b810} Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.258995 42179 prober.go:118] Liveness probe for "kube-apiserver-pike_kube-system(89228f8427ffc79800b43469845b379e):kube-apiserver" succeeded Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.271062 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.271171 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.371922 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.372429 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.473921 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.474034 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.500504 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 8081, Path: /readiness Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.500834 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.502088 42179 http.go:96] Probe succeeded for http://192.168.0.43:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:21 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc420f34a20 3 [] true false map[] 0xc4227d1000 } Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.502295 42179 prober.go:118] Readiness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.524616 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.524735 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.525853 42179 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Content-Type:[text/plain; charset=utf-8] Date:[Mon, 02 Jul 2018 20:40:21 GMT]] 0xc421a43c60 2 [] true false map[] 0xc42058b800 } Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.525950 42179 prober.go:118] Liveness probe for "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e):kube-scheduler" succeeded Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.574632 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.574783 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.676230 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.676383 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.700202 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.777756 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.777912 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.879535 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.879635 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.980354 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:21 pike kubelet[42179]: I0702 16:40:21.980409 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.081273 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.081331 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.181903 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.181997 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.282468 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.282572 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.383011 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.383075 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.483709 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.483767 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.584143 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.584233 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.684925 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.685014 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.786724 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.786873 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.888310 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.888819 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.989746 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:22 pike kubelet[42179]: I0702 16:40:22.989870 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.090412 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.090526 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.185369 42179 csi_attacher.go:362] kubernetes.io/csi: attacher.MountDevice successfully requested NodeStageVolume [/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.185431 42179 operation_generator.go:514] MountVolume.MountDevice succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") device mount path "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.185446 42179 operation_generator.go:591] Resizing is not enabled for this volume kubernetes.io/csi/io.netapp.trident.csi^{"name":"pvc-281709ce7e3711e8","protocol":"block"} Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.185464 42179 csi_mounter.go:77] kubernetes.io/csi: mounter.GetPath generated [/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.185469 42179 csi_mounter.go:98] kubernetes.io/csi: Mounter.SetUpAt(/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount) Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.185512 42179 csi_client.go:242] kubernetes.io/csi: calling NodeGetCapabilities rpc Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.188859 42179 csi_mounter.go:161] kubernetes.io/csi: created target path successfully [/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.188971 42179 csi_mounter.go:311] kubernetes.io/csi: volume data file saved successfully [/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/vol_data.json] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.188984 42179 csi_client.go:117] kubernetes.io/csi: calling NodePublishVolume rpc [volid={"name":"pvc-281709ce7e3711e8","protocol":"block"},target_path=/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.190822 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.190878 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.241669 42179 interface.go:360] Looking for default routes with IPv4 addresses Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.241690 42179 interface.go:365] Default route transits interface "ens33" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.242435 42179 interface.go:174] Interface ens33 is up Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.242494 42179 interface.go:222] Interface "ens33" has 2 addresses :[192.168.228.17/24 fe80::779f:8935:8254:8bd8/64]. Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.242509 42179 interface.go:189] Checking addr 192.168.228.17/24. Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.242516 42179 interface.go:196] IP found 192.168.228.17 Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.242521 42179 interface.go:228] Found valid IPv4 address 192.168.228.17 for interface "ens33". Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.242526 42179 interface.go:371] Found active IP 192.168.228.17 Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.291265 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.291394 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.378006 42179 csi_mounter.go:217] kubernetes.io/csi: mounter.SetUp successfully requested NodePublish [/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.378195 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.378266 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "pvc-281709ce7e3711e8" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.388764 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /readiness Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.388838 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.389926 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/readiness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:23 GMT]] 0xc4209c40e0 0 [] true false map[] 0xc422459600 } Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.389973 42179 prober.go:118] Readiness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.432264 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.432358 42179 kuberuntime_manager.go:385] No sandbox for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" can be found. Need to start a new one Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.432406 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.432489 42179 kuberuntime_manager.go:579] SyncPod received new pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)", will create a sandbox for it Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.432516 42179 kuberuntime_manager.go:588] Stopping PodSandbox for "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)", will start new one Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.432566 42179 kuberuntime_manager.go:640] Creating sandbox for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.434402 42179 docker_service.go:440] Setting cgroup parent to: "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.699937 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.792896 42179 factory.go:113] Using factory "docker" for container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.794842 42179 docker_sandbox.go:682] Will attempt to re-write config file /var/lib/docker/containers/366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2/resolv.conf with: Jul 02 16:40:23 pike kubelet[42179]: [nameserver 10.96.0.10 search trident.svc.cluster.local svc.cluster.local cluster.local localdomain options ndots:5] Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.794995 42179 plugins.go:412] Calling network plugin cni to set up pod "pod2_trident" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.796561 42179 manager.go:997] Added container: "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" (aliases: [k8s_POD_pod2_trident_18ea1a7c-7e38-11e8-a286-005056383598_0 366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2], namespace: "docker") Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.797019 42179 handler.go:325] Added event &{/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2 2018-07-02 20:40:23.43731516 +0000 UTC containerCreation {}} Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.797083 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.798902 42179 cni.go:286] Got netns path /proc/53153/ns/net Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.798941 42179 cni.go:287] Using podns path trident Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.799491 42179 cni.go:256] About to add CNI network cni-loopback (type=loopback) Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.805527 42179 cni.go:286] Got netns path /proc/53153/ns/net Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.805562 42179 cni.go:287] Using podns path trident Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.805784 42179 cni.go:256] About to add CNI network k8s-pod-network (type=calico) Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.863 [INFO][53202] calico.go 75: Extracted identifiers EndpointIDs=&utils.WEPIdentifiers{Namespace:"trident", WEPName:"", WorkloadEndpointIdentifiers:names.WorkloadEndpointIdentifiers{Node:"pike", Orchestrator:"k8s", Endpoint:"eth0", Workload:"", Pod:"pod2", ContainerID:"366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2"}} Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.863 [INFO][53202] calico.go 76: Loaded CNI NetConf NetConfg=types.NetConf{CNIVersion:"0.3.0", Name:"k8s-pod-network", Type:"calico", IPAM:struct { Name string; Type string "json:\"type\""; Subnet string "json:\"subnet\""; AssignIpv4 *string "json:\"assign_ipv4\""; AssignIpv6 *string "json:\"assign_ipv6\""; IPv4Pools []string "json:\"ipv4_pools,omitempty\""; IPv6Pools []string "json:\"ipv6_pools,omitempty\"" }{Name:"", Type:"host-local", Subnet:"usePodCidr", AssignIpv4:(*string)(nil), AssignIpv6:(*string)(nil), IPv4Pools:[]string(nil), IPv6Pools:[]string(nil)}, Args:types.Args{Mesos:types.Mesos{NetworkInfo:types.NetworkInfo{Name:"", Labels:struct { Labels []struct { Key string "json:\"key\""; Value string "json:\"value\"" } "json:\"labels,omitempty\"" }{Labels:[]struct { Key string "json:\"key\""; Value string "json:\"value\"" }(nil)}}}}, MTU:1500, Nodename:"pike", NodenameFileOptional:false, DatastoreType:"kubernetes", EtcdEndpoints:"", LogLevel:"info", Policy:types.Policy{PolicyType:"k8s", K8sAPIRoot:"", K8sAuthToken:"", K8sClientCertificate:"", K8sClientKey:"", K8sCertificateAuthority:""}, Kubernetes:types.Kubernetes{K8sAPIRoot:"", Kubeconfig:"/etc/cni/net.d/calico-kubeconfig", NodeName:""}, EtcdScheme:"", EtcdKeyFile:"", EtcdCertFile:"", EtcdCaCertFile:"", ContainerSettings:types.ContainerSettings{AllowIPForwarding:false}, EtcdAuthority:"", Hostname:""} Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.863 [INFO][53202] utils.go 379: Configured environment: [CNI_COMMAND=ADD CNI_CONTAINERID=366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2 CNI_NETNS=/proc/53153/ns/net CNI_ARGS=IgnoreUnknown=1;IgnoreUnknown=1;K8S_POD_NAMESPACE=trident;K8S_POD_NAME=pod2;K8S_POD_INFRA_CONTAINER_ID=366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2 CNI_IFNAME=eth0 CNI_PATH=/opt/calico/bin:/opt/cni/bin LANG=en_US.UTF-8 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf KUBELET_SYSTEM_PODS_ARGS=--pod-manifest-path=/etc/kubernetes/manifests --allow-privileged=true KUBELET_NETWORK_ARGS=--network-plugin=cni --cni-conf-dir=/etc/cni/net.d --cni-bin-dir=/opt/cni/bin KUBELET_DNS_ARGS=--cluster-dns=10.96.0.10 --cluster-domain=cluster.local KUBELET_AUTHZ_ARGS=--authorization-mode=Webhook --client-ca-file=/etc/kubernetes/pki/ca.crt KUBELET_CADVISOR_ARGS=--cadvisor-port=0 KUBELET_CERTIFICATE_ARGS=--rotate-certificates=true --cert-dir=/var/lib/kubelet/pki DATASTORE_TYPE=kubernetes KUBECONFIG=/etc/cni/net.d/calico-kubeconfig] Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.914 [INFO][53202] calico.go 167: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {pike-k8s-pod2-eth0 trident 18ea1a7c-7e38-11e8-a286-005056383598 149636 0 2018-07-02 16:40:03 -0400 EDT map[projectcalico.org/namespace:trident projectcalico.org/orchestrator:k8s] map[] [] nil [] } {k8s pike pod2 eth0 [] [] [kns.trident] caliccaeca26c1e []}} ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-" Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.915 [INFO][53202] k8s.go 60: Extracted identifiers for CmdAddK8s ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" Jul 02 16:40:23 pike kubelet[42179]: Calico CNI fetching podCidr from Kubernetes Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.927 [INFO][53202] k8s.go 83: Fetched podCidr ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" podCidr="192.168.0.0/24" Jul 02 16:40:23 pike kubelet[42179]: Calico CNI passing podCidr to host-local IPAM: 192.168.0.0/24 Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.930 [INFO][53202] k8s.go 660: pod info &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:pod2,GenerateName:,Namespace:trident,SelfLink:/api/v1/namespaces/trident/pods/pod2,UID:18ea1a7c-7e38-11e8-a286-005056383598,ResourceVersion:149636,Generation:0,CreationTimestamp:2018-07-02 16:40:03 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{data {nil nil nil nil nil nil nil nil nil PersistentVolumeClaimVolumeSource{ClaimName:pvc2,ReadOnly:false,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {default-token-9dv6w {nil nil nil nil nil &SecretVolumeSource{SecretName:default-token-9dv6w,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:latest [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{data false /var/www } {default-token-9dv6w true /var/run/secrets/kubernetes.io/serviceaccount }] nil nil nil /dev/termination-log File Always nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:pike,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc42031ee90} {node.kubernetes.io/unreachable Exists NoExecute 0xc42031eeb0}],HostAliases:[],PriorityClassName:,Priority:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:40:03 -0400 EDT } {Ready False 0001-01-01 00:0 Jul 02 16:40:23 pike kubelet[42179]: 0:00 +0000 UTC 2018-07-02 16:40:03 -0400 EDT ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:40:03 -0400 EDT }],Message:,Reason:,HostIP:192.168.228.17,PodIP:,StartTime:2018-07-02 16:40:03 -0400 EDT,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:latest }],QOSClass:BestEffort,InitContainerStatuses:[],},} Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.933 [INFO][53202] k8s.go 267: Populated endpoint ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"pike-k8s-pod2-eth0", GenerateName:"", Namespace:"trident", SelfLink:"", UID:"18ea1a7c-7e38-11e8-a286-005056383598", ResourceVersion:"149636", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666160803, loc:(*time.Location)(0x1ec6320)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/orchestrator":"k8s", "projectcalico.org/namespace":"trident"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"pike", ContainerID:"", Pod:"pod2", Endpoint:"eth0", IPNetworks:[]string{"192.168.0.53/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.trident"}, InterfaceName:"caliccaeca26c1e", MAC:"", Ports:[]v3.EndpointPort(nil)}} Jul 02 16:40:23 pike kubelet[42179]: Calico CNI using IPs: [192.168.0.53/32] Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.934 [INFO][53202] network.go 31: Setting the host side veth name to caliccaeca26c1e ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.935 [INFO][53202] network.go 326: Disabling IPv4 forwarding ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.967 [INFO][53202] k8s.go 294: Added Mac, interface name, and active container ID to endpoint ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"pike-k8s-pod2-eth0", GenerateName:"", Namespace:"trident", SelfLink:"", UID:"18ea1a7c-7e38-11e8-a286-005056383598", ResourceVersion:"149636", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666160803, loc:(*time.Location)(0x1ec6320)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"trident", "projectcalico.org/orchestrator":"k8s"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"pike", ContainerID:"366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2", Pod:"pod2", Endpoint:"eth0", IPNetworks:[]string{"192.168.0.53/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.trident"}, InterfaceName:"caliccaeca26c1e", MAC:"96:94:2a:28:a8:29", Ports:[]v3.EndpointPort(nil)}} Jul 02 16:40:23 pike kubelet[42179]: 2018-07-02 16:40:23.983 [INFO][53202] k8s.go 302: Wrote updated endpoint to datastore ContainerID="366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" Namespace="trident" Pod="pod2" WorkloadEndpoint="pike-k8s-pod2-eth0" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.985092 42179 config.go:297] Setting pods for source api Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.985768 42179 kubelet.go:1863] SyncLoop (UPDATE, "api"): "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:23 pike kubelet[42179]: I0702 16:40:23.995190 42179 kuberuntime_manager.go:654] Created PodSandbox "366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2" for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.000533 42179 kuberuntime_manager.go:673] Determined the ip "192.168.0.53" for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" after sandbox changed Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.000674 42179 kuberuntime_manager.go:724] Creating container &Container{Name:nginx,Image:nginx:latest,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TCP }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{data false /var/www } {default-token-9dv6w true /var/run/secrets/kubernetes.io/serviceaccount }],LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,VolumeDevices:[],} in pod pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598) Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.002752 42179 kuberuntime_image.go:47] Pulling image "nginx:latest" without credentials Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.002784 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx:latest" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.144648 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /liveness Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.144757 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.146009 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/liveness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:24 GMT]] 0xc42296a880 0 [] true false map[] 0xc421202800 } Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.146291 42179 prober.go:118] Liveness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.380556 42179 generic.go:147] GenericPLEG: 18ea1a7c-7e38-11e8-a286-005056383598/366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2: non-existent -> running Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.382254 42179 kuberuntime_manager.go:852] getSandboxIDByPodUID got sandbox IDs ["366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2"] for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.386789 42179 generic.go:380] PLEG: Write status for pod2/trident: &container.PodStatus{ID:"18ea1a7c-7e38-11e8-a286-005056383598", Name:"pod2", Namespace:"trident", IP:"192.168.0.53", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*v1alpha2.PodSandboxStatus{(*v1alpha2.PodSandboxStatus)(0xc4227d39f0)}} (err: ) Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.386829 42179 kubelet.go:1901] SyncLoop (PLEG): "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)", event: &pleg.PodLifecycleEvent{ID:"18ea1a7c-7e38-11e8-a286-005056383598", Type:"ContainerStarted", Data:"366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2"} Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.523193 42179 kube_docker_client.go:348] Stop pulling image "nginx:latest": "Status: Image is up to date for nginx:latest" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.526266 42179 kuberuntime_container.go:100] Generating ref for container nginx: &v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:"spec.containers{nginx}"} Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.526610 42179 kubelet_pods.go:172] container: trident/pod2/nginx podIP: "192.168.0.53" creating hosts mount: true Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.526854 42179 csi_mounter.go:77] kubernetes.io/csi: mounter.GetPath generated [/var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.526804 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "nginx:latest" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.530277 42179 expiration_cache.go:98] Entry version: {key:version obj:0xc421b925a0} has expired Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.531693 42179 docker_service.go:440] Setting cgroup parent to: "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.614169 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /metrics Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.614224 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.620846 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; version=0.0.4] Date:[Mon, 02 Jul 2018 20:40:24 GMT]] 0xc422a83a20 -1 [] true true map[] 0xc420a9af00 } Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.620907 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):sidecar" succeeded Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.640040 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Created' Created container Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.676479 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.811956 42179 factory.go:113] Using factory "docker" for container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f" Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.813256 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2", UID:"18ea1a7c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149635", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Started' Started container Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.814566 42179 manager.go:997] Added container: "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f" (aliases: [k8s_nginx_pod2_trident_18ea1a7c-7e38-11e8-a286-005056383598_0 5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f], namespace: "docker") Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.814697 42179 handler.go:325] Added event &{/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f 2018-07-02 20:40:24.53461108 +0000 UTC containerCreation {}} Jul 02 16:40:24 pike kubelet[42179]: I0702 16:40:24.814727 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod18ea1a7c-7e38-11e8-a286-005056383598/5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f" Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.400870 42179 generic.go:147] GenericPLEG: 18ea1a7c-7e38-11e8-a286-005056383598/5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f: non-existent -> running Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.402873 42179 kuberuntime_manager.go:852] getSandboxIDByPodUID got sandbox IDs ["366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2"] for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.410513 42179 generic.go:380] PLEG: Write status for pod2/trident: &container.PodStatus{ID:"18ea1a7c-7e38-11e8-a286-005056383598", Name:"pod2", Namespace:"trident", IP:"192.168.0.53", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc4214e6620)}, SandboxStatuses:[]*v1alpha2.PodSandboxStatus{(*v1alpha2.PodSandboxStatus)(0xc422f1b5e0)}} (err: ) Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.410576 42179 kubelet.go:1901] SyncLoop (PLEG): "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)", event: &pleg.PodLifecycleEvent{ID:"18ea1a7c-7e38-11e8-a286-005056383598", Type:"ContainerStarted", Data:"5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f"} Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.410609 42179 kubelet_pods.go:1366] Generating status for "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.410763 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.416222 42179 config.go:297] Setting pods for source api Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.416899 42179 kubelet.go:1869] SyncLoop (RECONCILE, "api"): "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.417186 42179 status_manager.go:481] Status for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:25 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.53 StartTime:2018-07-02 16:40:03 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:40:24 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:nginx:latest ImageID:docker-pullable://nginx@sha256:62a095e5da5f977b9f830adaf64d604c614024bf239d21068e4ca826d0d629a4 ContainerID:docker://5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f}] QOSClass:BestEffort}) Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.462827 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "18ea1a7c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.462899 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "18ea1a7c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.504114 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.504341 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 18ea1a7c-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.506946 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.507129 42179 atomic_writer.go:156] pod trident/pod2 volume default-token-9dv6w: no update required for target directory /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.507159 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.654470 42179 reflector.go:428] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Watch close - *v1.Pod total 29 items received Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.700364 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.711048 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:25 pike kubelet[42179]: I0702 16:40:25.711154 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.420011 42179 kubelet_pods.go:1366] Generating status for "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.420163 42179 status_manager.go:340] Ignoring same status for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:25 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:03 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.53 StartTime:2018-07-02 16:40:03 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:40:24 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:nginx:latest ImageID:docker-pullable://nginx@sha256:62a095e5da5f977b9f830adaf64d604c614024bf239d21068e4ca826d0d629a4 ContainerID:docker://5fe4004d45b0ff0688e57d81a594f416620c01128091eef0fe46b305c1bbf46f}] QOSClass:BestEffort} Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.420306 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.475562 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "18ea1a7c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.475642 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "18ea1a7c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.509305 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.509438 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 18ea1a7c-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.511074 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.511203 42179 atomic_writer.go:156] pod trident/pod2 volume default-token-9dv6w: no update required for target directory /var/lib/kubelet/pods/18ea1a7c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.511222 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/18ea1a7c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2" (UID: "18ea1a7c-7e38-11e8-a286-005056383598") Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.720699 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:26 pike kubelet[42179]: I0702 16:40:26.720853 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "pod2_trident(18ea1a7c-7e38-11e8-a286-005056383598)" Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.030449 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.030492 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.031715 42179 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421153980 2 [] true false map[] 0xc4224dfd00 } Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.031786 42179 prober.go:118] Liveness probe for "kube-controller-manager-pike_kube-system(fac56382eca21a00fc5c8a039563b311):kube-controller-manager" succeeded Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.212842 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:trident-csi-0,GenerateName:trident-csi-,Namespace:trident,SelfLink:/api/v1/namespaces/trident/pods/trident-csi-0,UID:8eaa19b4-7e34-11e8-a286-005056383598,ResourceVersion:147708,Generation:0,CreationTimestamp:2018-07-02 16:14:43 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{app: controller.csi.trident.netapp.io,controller-revision-hash: trident-csi-66f8f9d5bb,statefulset.kubernetes.io/pod-name: trident-csi-0,},Annotations:map[string]string{cni.projectcalico.org/podIP: 192.168.0.49/32,kubernetes.io/config.seen: 2018-07-02T16:23:31.650583191-04:00,kubernetes.io/config.source: api,},OwnerReferences:[{apps/v1 StatefulSet trident-csi 8e98152e-7e34-11e8-a286-005056383598 0xc420438d08 0xc420438d09}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-vol {nil nil nil nil nil nil nil nil nil PersistentVolumeClaimVolumeSource{ClaimName:trident-csi,ReadOnly:false,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {socket-dir {nil &EmptyDirVolumeSource{Medium:,SizeLimit:,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etc-dir {&HostPathVolumeSource{Path:/etc,Type:*Directory,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {trident-csi-token-4gzh4 {nil nil nil nil nil &SecretVolumeSource{SecretName:trident-csi-token-4gzh4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{trident-main registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae [/usr/local/bin/trident_orchestrator] [-etcd_v3 http://127.0.0.1:8001 --csi_node_name=$(KUBE_NODE_NAME) --csi_endpoint=$(CSI_ENDPOINT) -debug] [] [] [{KUBE_NODE_NAME EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldP Jul 02 16:40:27 pike kubelet[42179]: ath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,}} {CSI_ENDPOINT unix://plugin/csi.sock nil}] {map[] map[]} [{socket-dir false /plugin } {etc-dir false /etc } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[tridentctl -s 127.0.0.1:8000 get backend],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:120,TimeoutSeconds:90,PeriodSeconds:120,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false} {etcd quay.io/coreos/etcd:v3.2.19 [/usr/local/bin/etcd] [-name etcd1 -advertise-client-urls http://127.0.0.1:8001 -listen-client-urls http://127.0.0.1:8001 -initial-advertise-peer-urls http://127.0.0.1:8002 -listen-peer-urls http://127.0.0.1:8002 -data-dir /var/etcd/data -initial-cluster etcd1=http://127.0.0.1:8002] [] [] [] {map[] map[]} [{etcd-vol false /var/etcd/data } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:10,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false} {csi-attacher quay.io/k8scsi/csi-attacher:v0.2.0 [] [--v=9 --csi-address=$(ADDRESS)] [] [] [{ADDRESS /var/lib/csi/sockets/pluginproxy/csi.sock nil}] {map[] map[]} [{socket-dir false /var/lib/csi/sockets/pluginproxy/ } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false} {csi-provisioner quay.io/k8scsi/csi-provisioner:v0.2.1 [] [--v=9 --provisioner=io.netapp.trident.csi --csi-address=$(ADDRESS)] [] [] [{ADDRESS /var/lib/csi/sockets/pluginproxy/csi.sock nil}] {map[] map[]} [{socket-dir false /var/lib/csi/sockets/pluginproxy/ } {trident-csi-token-4gzh4 true /var/run/s Jul 02 16:40:27 pike kubelet[42179]: ecrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:trident-csi,DeprecatedServiceAccount:trident-csi,NodeName:pike,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:trident-csi-0,Subdomain:trident-csi,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc420438d18} {node.kubernetes.io/unreachable Exists NoExecute 0xc420438d20}],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:43 -0400 EDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:48 -0400 EDT } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:43 -0400 EDT }],Message:,Reason:,HostIP:192.168.228.17,PodIP:192.168.0.49,StartTime:2018-07-02 16:14:43 -0400 EDT,ContainerStatuses:[{csi-attacher {nil ContainerStateRunning{StartedAt:2018-07-02 16:14:47 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/k8scsi/csi-attacher:v0.2.0 docker-pullable://quay.io/k8scsi/csi-attacher@sha256:5cbb7934bd86d400c221379cff8b24ed4c06e121ea59608cfd7e67690100ba54 docker://f547da90c65d01207a828ee3b8d044eb4b4cc05fa065f39648eac3e539f3b9a6} {csi-provisioner {nil &ContainerStateRunning{StartedAt:2018-07-02 16:14:48 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/k8scsi/csi-provisioner:v0.2.1 docker-pullable://quay.io/k8scsi/csi-provisioner@sha256:fd4ed32315e846b6654f97c95b373da001fd9638cd5935c20a5bf9f5889e8602 docker://08d31d6c565ec486d8e9c46017c10edd0f390da6a139692a9daa25e620777bc0} {etcd {nil &ContainerStateRunning{StartedAt:2018-0 Jul 02 16:40:27 pike kubelet[42179]: 7-02 16:14:47 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/coreos/etcd:v3.2.19 docker-pullable://quay.io/coreos/etcd@sha256:e21b9734136122b1e8b9b37742e72d2bd8b0d84bca875892494e0de5520313b0 docker://8340342e4a7288ef0672e62be7c0bf34e1800f7ce961607fa163146aa828b861} {trident-main {nil &ContainerStateRunning{StartedAt:2018-07-02 16:14:46 -0400 EDT,} nil} {nil nil nil} true 0 registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae docker-pullable://registry.barnacle.netapp.com:5000/trident@sha256:ce8cd402f9ef85e18ce428f09137a7347c5009c255daf730f9c40c006ea0ee23 docker://2c51e41d415063f9be931cf9ebbb62bf4d4c427d9002f312fbf54f9c34a4d9ee}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd quay.io/coreos/etcd:v3.2.19 [/usr/local/bin/etcd] [-name etcd1 -advertise-client-urls http://127.0.0.1:8001 -listen-client-urls http://127.0.0.1:8001 -initial-advertise-peer-urls http://127.0.0.1:8002 -listen-peer-urls http://127.0.0.1:8002 -data-dir /var/etcd/data -initial-cluster etcd1=http://127.0.0.1:8002] [] [] [] {map[] map[]} [{etcd-vol false /var/etcd/data } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:10,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health] Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.280755 42179 exec.go:38] Exec probe response: "member 9ca0c7938a0c3dc5 is healthy: got healthy result from http://127.0.0.1:8001\ncluster is healthy\n" Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.280848 42179 prober.go:118] Liveness probe for "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598):etcd" succeeded Jul 02 16:40:27 pike kubelet[42179]: I0702 16:40:27.700459 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.700158 42179 kubelet.go:1920] SyncLoop (SYNC): 1 pods; kube-proxy-2p4fh_kube-system(baeff5dc-773a-11e8-a95f-005056383598) Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.700336 42179 kubelet_pods.go:1366] Generating status for "kube-proxy-2p4fh_kube-system(baeff5dc-773a-11e8-a95f-005056383598)" Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.700527 42179 status_manager.go:340] Ignoring same status for pod "kube-proxy-2p4fh_kube-system(baeff5dc-773a-11e8-a95f-005056383598)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-23 19:11:16 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-29 12:55:57 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-23 19:11:16 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.228.17 StartTime:2018-06-23 19:11:16 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:kube-proxy State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-06-29 12:55:30 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:2,Signal:0,Reason:Error,Message:,StartedAt:2018-06-29 12:50:29 -0400 EDT,FinishedAt:2018-06-29 12:55:10 -0400 EDT,ContainerID:docker://c9e2ee500250171bc761b0aef67b4480a683bc1de37117701397e140538708d4,}} Ready:true RestartCount:3 Image:k8s.gcr.io/kube-proxy-amd64:v1.10.5 ImageID:docker-pullable://k8s.gcr.io/kube-proxy-amd64@sha256:b81228e8ad694f05a5a6e035167ad705600aead5cfd63628a38984fc60f0b989 ContainerID:docker://5967cea5631752f4d86900e5f7c3a179141a91b0879c7850b7d177a32987afa3}] QOSClass:BestEffort} Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.700830 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "kube-proxy-2p4fh_kube-system(baeff5dc-773a-11e8-a95f-005056383598)" Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.790589 42179 desired_state_of_world_populator.go:302] Added volume "kube-proxy" (volSpec="kube-proxy") for pod "baeff5dc-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.790639 42179 desired_state_of_world_populator.go:302] Added volume "xtables-lock" (volSpec="xtables-lock") for pod "baeff5dc-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.790652 42179 desired_state_of_world_populator.go:302] Added volume "lib-modules" (volSpec="lib-modules") for pod "baeff5dc-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.790689 42179 desired_state_of_world_populator.go:302] Added volume "kube-proxy-token-q4qcr" (volSpec="kube-proxy-token-q4qcr") for pod "baeff5dc-773a-11e8-a95f-005056383598" to desired state. Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.828473 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "kube-proxy-token-q4qcr" (UniqueName: "kubernetes.io/secret/baeff5dc-773a-11e8-a95f-005056383598-kube-proxy-token-q4qcr") pod "kube-proxy-2p4fh" (UID: "baeff5dc-773a-11e8-a95f-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.829121 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/baeff5dc-773a-11e8-a95f-005056383598-kube-proxy") pod "kube-proxy-2p4fh" (UID: "baeff5dc-773a-11e8-a95f-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.829348 42179 secret.go:186] Setting up volume kube-proxy-token-q4qcr for pod baeff5dc-773a-11e8-a95f-005056383598 at /var/lib/kubelet/pods/baeff5dc-773a-11e8-a95f-005056383598/volumes/kubernetes.io~secret/kube-proxy-token-q4qcr Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.829630 42179 configmap.go:187] Setting up volume kube-proxy for pod baeff5dc-773a-11e8-a95f-005056383598 at /var/lib/kubelet/pods/baeff5dc-773a-11e8-a95f-005056383598/volumes/kubernetes.io~configmap/kube-proxy Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.831738 42179 secret.go:216] Received secret kube-system/kube-proxy-token-q4qcr containing (3) pieces of data, 1900 total bytes Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.831881 42179 atomic_writer.go:156] pod kube-system/kube-proxy-2p4fh volume kube-proxy-token-q4qcr: no update required for target directory /var/lib/kubelet/pods/baeff5dc-773a-11e8-a95f-005056383598/volumes/kubernetes.io~secret/kube-proxy-token-q4qcr Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.831904 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "kube-proxy-token-q4qcr" (UniqueName: "kubernetes.io/secret/baeff5dc-773a-11e8-a95f-005056383598-kube-proxy-token-q4qcr") pod "kube-proxy-2p4fh" (UID: "baeff5dc-773a-11e8-a95f-005056383598") Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.831987 42179 configmap.go:217] Received configMap kube-system/kube-proxy containing (2) pieces of data, 1220 total bytes Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.832281 42179 atomic_writer.go:156] pod kube-system/kube-proxy-2p4fh volume kube-proxy: no update required for target directory /var/lib/kubelet/pods/baeff5dc-773a-11e8-a95f-005056383598/volumes/kubernetes.io~configmap/kube-proxy Jul 02 16:40:28 pike kubelet[42179]: I0702 16:40:28.832306 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/baeff5dc-773a-11e8-a95f-005056383598-kube-proxy") pod "kube-proxy-2p4fh" (UID: "baeff5dc-773a-11e8-a95f-005056383598") Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.001287 42179 volume_manager.go:380] All volumes are attached and mounted for pod "kube-proxy-2p4fh_kube-system(baeff5dc-773a-11e8-a95f-005056383598)" Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.001965 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:f0bc7d737878c4ededb7e627c3ba4ef3480dbb6abd608793e0d754d3812159ae Attempt:3 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-proxy-2p4fh_kube-system(baeff5dc-773a-11e8-a95f-005056383598)" Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.207492 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/dnsmasq Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.207616 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.209018 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:29 GMT] Content-Length:[51]] 0xc422b42760 51 [] true false map[] 0xc4228c2300 } Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.209415 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):dnsmasq" succeeded Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.678926 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:29 pike kubelet[42179]: I0702 16:40:29.700200 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.115988 42179 eviction_manager.go:229] eviction manager: synchronize housekeeping Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.178636 42179 helpers.go:840] eviction manager: observations: signal=imagefs.inodesFree, available: 1687910, capacity: 2240Ki, time: 2018-07-02 16:40:30.136071915 -0400 EDT m=+1018.669761288 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.178932 42179 helpers.go:840] eviction manager: observations: signal=pid.available, available: 129772, capacity: 128Ki, time: 2018-07-02 16:40:30.174668881 -0400 EDT m=+1018.708358230 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.179108 42179 helpers.go:840] eviction manager: observations: signal=memory.available, available: 1543748Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:30.136071915 -0400 EDT m=+1018.669761288 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.179278 42179 helpers.go:840] eviction manager: observations: signal=allocatableMemory.available, available: 5871372Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:30.178487482 -0400 EDT m=+1018.712176854 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.179461 42179 helpers.go:840] eviction manager: observations: signal=nodefs.available, available: 10752516Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:30.136071915 -0400 EDT m=+1018.669761288 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.179588 42179 helpers.go:840] eviction manager: observations: signal=nodefs.inodesFree, available: 1687910, capacity: 2240Ki, time: 2018-07-02 16:40:30.136071915 -0400 EDT m=+1018.669761288 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.179729 42179 helpers.go:840] eviction manager: observations: signal=imagefs.available, available: 10752516Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:30.136071915 -0400 EDT m=+1018.669761288 Jul 02 16:40:30 pike kubelet[42179]: I0702 16:40:30.179916 42179 eviction_manager.go:336] eviction manager: no resources are starved Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.044882 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/kubedns Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.044926 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.046108 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:31 GMT] Content-Length:[49]] 0xc422abe880 49 [] true false map[] 0xc4203b5800 } Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.046197 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.163362 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:etcd-pike,GenerateName:,Namespace:kube-system,SelfLink:/api/v1/namespaces/kube-system/pods/etcd-pike,UID:68f9c58845a7c8f7ac9d6261d498ab6c,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{component: etcd,tier: control-plane,},Annotations:map[string]string{kubernetes.io/config.hash: 68f9c58845a7c8f7ac9d6261d498ab6c,kubernetes.io/config.seen: 2018-07-02T16:23:31.6928551-04:00,kubernetes.io/config.source: file,scheduler.alpha.kubernetes.io/critical-pod: ,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-data {HostPathVolumeSource{Path:/var/lib/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etcd-certs {&HostPathVolumeSource{Path:/etc/kubernetes/pki/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/h Jul 02 16:40:31 pike kubelet[42179]: ealthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:pike,HostNetwork:true,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{ Exists NoExecute }],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds: Jul 02 16:40:31 pike kubelet[42179]: 10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo] Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.238447 42179 exec.go:38] Exec probe response: "2018-07-02 20:40:31.208950 I | warning: ignoring ServerName for user-provided CA for backwards compatibility is deprecated\n" Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.238469 42179 prober.go:118] Liveness probe for "etcd-pike_kube-system(68f9c58845a7c8f7ac9d6261d498ab6c):etcd" succeeded Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.252111 42179 prober.go:165] HTTP-Probe Host: https://192.168.228.17, Port: 6443, Path: /healthz Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.252172 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.257554 42179 http.go:96] Probe succeeded for https://192.168.228.17:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Mon, 02 Jul 2018 20:40:31 GMT]] 0xc420b69de0 2 [] false false map[] 0xc42058a200 0xc4208eec60} Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.257597 42179 prober.go:118] Liveness probe for "kube-apiserver-pike_kube-system(89228f8427ffc79800b43469845b379e):kube-apiserver" succeeded Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.499741 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 8081, Path: /readiness Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.499813 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.500671 42179 http.go:96] Probe succeeded for http://192.168.0.43:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:31 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc4217c5700 3 [] true false map[] 0xc42185dc00 } Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.500730 42179 prober.go:118] Readiness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.524950 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.525021 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.525889 42179 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:31 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc422e64b20 2 [] true false map[] 0xc42185de00 } Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.525957 42179 prober.go:118] Liveness probe for "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e):kube-scheduler" succeeded Jul 02 16:40:31 pike kubelet[42179]: I0702 16:40:31.700047 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.113898 42179 kubelet.go:1265] Container garbage collection succeeded Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.614097 42179 config.go:297] Setting pods for source api Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.614513 42179 config.go:405] Receiving a new pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.614571 42179 kubelet.go:1856] SyncLoop (ADD, "api"): "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.622713 42179 kubelet_pods.go:1366] Generating status for "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.623096 42179 qos_container_manager_linux.go:317] [ContainerManager]: Updated QoS cgroup configuration Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.628247 42179 factory.go:117] Factory "docker" was unable to handle container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.628379 42179 factory.go:106] Error trying to work out if we can handle /kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598: /kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598 not handled by systemd handler Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.628388 42179 factory.go:117] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.628394 42179 factory.go:113] Using factory "raw" for container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.628846 42179 manager.go:997] Added container: "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" (aliases: [], namespace: "") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.629351 42179 handler.go:325] Added event &{/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598 2018-07-02 16:40:32.621789796 -0400 EDT containerCreation {}} Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.629495 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.630445 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.637617 42179 status_manager.go:481] Status for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" updated successfully: (1, {Phase:Pending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP: StartTime:2018-07-02 16:40:32 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:nginx:latest ImageID: ContainerID:}] QOSClass:BestEffort}) Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.638944 42179 config.go:297] Setting pods for source api Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.639423 42179 kubelet.go:1869] SyncLoop (RECONCILE, "api"): "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.718614 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "2a3383f3-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.718689 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "2a3383f3-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.756231 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.756743 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.756956 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.757019 42179 reconciler.go:252] operationExecutor.MountVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.757079 42179 operation_generator.go:486] MountVolume.WaitForAttach entering for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") DevicePath "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.757089 42179 csi_attacher.go:131] kubernetes.io/csi: probing for updates from CSI driver for [attachment.ID=csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c] Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.757110 42179 csi_attacher.go:140] kubernetes.io/csi: probing VolumeAttachment [id=csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c] Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.760732 42179 operation_generator.go:495] MountVolume.WaitForAttach succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") DevicePath "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.760916 42179 csi_attacher.go:244] kubernetes.io/csi: attacher.GetDeviceMountPath(&{nil &PersistentVolume{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:pvc-281709ce7e3711e8,GenerateName:,Namespace:,SelfLink:/api/v1/persistentvolumes/pvc-281709ce7e3711e8,UID:2c19eed1-7e37-11e8-a286-005056383598,ResourceVersion:149392,Generation:0,CreationTimestamp:2018-07-02 16:33:26 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{pv.kubernetes.io/provisioned-by: io.netapp.trident.csi,},OwnerReferences:[],Finalizers:[kubernetes.io/pv-protection external-attacher/io-netapp-trident-csi],ClusterName:,Initializers:nil,},Spec:PersistentVolumeSpec{Capacity:ResourceList{storage: {{1073741824 0} {} 1Gi BinarySI},},PersistentVolumeSource:PersistentVolumeSource{GCEPersistentDisk:nil,AWSElasticBlockStore:nil,HostPath:nil,Glusterfs:nil,NFS:nil,RBD:nil,ISCSI:nil,Cinder:nil,CephFS:nil,FC:nil,Flocker:nil,FlexVolume:nil,AzureFile:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Local:nil,StorageOS:nil,CSI:&CSIPersistentVolumeSource{Driver:io.netapp.trident.csi,VolumeHandle:{"name":"pvc-281709ce7e3711e8","protocol":"block"},ReadOnly:false,FSType:,VolumeAttributes:map[string]string{backend: solidfire_10.250.118.144,internalName: pvc-281709ce7e3711e8,name: pvc-281709ce7e3711e8,protocol: block,storage.kubernetes.io/csiProvisionerIdentity: 1530562488326-8081-io.netapp.trident.csi,},ControllerPublishSecretRef:nil,NodeStageSecretRef:nil,NodePublishSecretRef:nil,},},AccessModes:[ReadWriteMany],ClaimRef:&ObjectReference{Kind:PersistentVolumeClaim,Namespace:trident,Name:pvc2,UID:281709ce-7e37-11e8-a286-005056383598,APIVersion:v1,ResourceVersion:149109,FieldPath:,},PersistentVolumeReclaimPolicy:Delete,StorageClassName:ssd-sf,MountOptions:[],VolumeMode:nil,NodeAffinity:nil,},Status:PersistentVolumeStatus{Phase:Bound,Message:,Reason:,},} false}) Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.760939 42179 csi_attacher.go:250] attacher.GetDeviceMountPath succeeded, deviceMountPath: /var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.760947 42179 csi_attacher.go:255] kubernetes.io/csi: attacher.MountDevice(csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c, /var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount) Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.760975 42179 csi_client.go:242] kubernetes.io/csi: calling NodeGetCapabilities rpc Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.764803 42179 csi_attacher.go:326] kubernetes.io/csi: created target path successfully [/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.764841 42179 csi_client.go:186] kubernetes.io/csi: calling NodeStageVolume rpc [volid={"name":"pvc-281709ce7e3711e8","protocol":"block"},staging_target_path=/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858283 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858507 42179 reconciler.go:252] operationExecutor.MountVolume started for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858549 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858587 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858674 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 2a3383f3-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858858 42179 empty_dir.go:256] pod 2a3383f3-7e38-11e8-a286-005056383598: mounting tmpfs for volume wrapped_default-token-9dv6w Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.858881 42179 mount_linux.go:143] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w]) Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.863543 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.863621 42179 factory.go:106] Error trying to work out if we can handle /system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope: /system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope not handled by systemd handler Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.863629 42179 factory.go:117] Factory "systemd" was unable to handle container "/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.863635 42179 factory.go:113] Using factory "raw" for container "/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.863958 42179 manager.go:997] Added container: "/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope" (aliases: [], namespace: "") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.864214 42179 handler.go:325] Added event &{/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope 2018-07-02 16:40:32.857779123 -0400 EDT containerCreation {}} Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.864265 42179 container.go:448] Start housekeeping for container "/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.865572 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-7b81eb46dd0fac833eff1ea683b75ba967ebb433b856dbf236825a8fdb8f502b.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.865592 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-7b81eb46dd0fac833eff1ea683b75ba967ebb433b856dbf236825a8fdb8f502b.mount", but ignoring. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.865600 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-7b81eb46dd0fac833eff1ea683b75ba967ebb433b856dbf236825a8fdb8f502b.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.871866 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-f32b41b009887b275502f2590c2a03f04a1aca9e87f56d650864c78ac73c464a.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.871894 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-f32b41b009887b275502f2590c2a03f04a1aca9e87f56d650864c78ac73c464a.mount", but ignoring. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.871907 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-f32b41b009887b275502f2590c2a03f04a1aca9e87f56d650864c78ac73c464a.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.872260 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-ee3b33397c0d.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.872274 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-docker-netns-ee3b33397c0d.mount", but ignoring. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.872280 42179 manager.go:957] ignoring container "/system.slice/run-docker-netns-ee3b33397c0d.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.874746 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.874765 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount", but ignoring. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.874793 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.877562 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.877594 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount", but ignoring. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.877608 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.878897 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2-shm.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.878920 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2-shm.mount", but ignoring. Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.878932 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2-shm.mount" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.879108 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.879245 42179 atomic_writer.go:176] pod trident/pod2b volume default-token-9dv6w: performed write of new data to ts data directory: /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w/..2018_07_02_20_40_32.018150428 Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.879355 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.879554 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-9dv6w" Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.880724 42179 manager.go:1054] Destroyed container: "/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope" (aliases: [], namespace: "") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.881290 42179 handler.go:325] Added event &{/system.slice/run-rd54da0887fea4f6b8d5ca6b7670a4c66.scope 2018-07-02 16:40:32.881278502 -0400 EDT m=+1021.414967893 containerDeletion {}} Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.959483 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:32 pike kubelet[42179]: I0702 16:40:32.959554 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.060341 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.060462 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.161090 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.161774 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.256374 42179 interface.go:360] Looking for default routes with IPv4 addresses Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.256413 42179 interface.go:365] Default route transits interface "ens33" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.257438 42179 interface.go:174] Interface ens33 is up Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.257527 42179 interface.go:222] Interface "ens33" has 2 addresses :[192.168.228.17/24 fe80::779f:8935:8254:8bd8/64]. Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.257544 42179 interface.go:189] Checking addr 192.168.228.17/24. Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.257551 42179 interface.go:196] IP found 192.168.228.17 Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.257557 42179 interface.go:228] Found valid IPv4 address 192.168.228.17 for interface "ens33". Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.257562 42179 interface.go:371] Found active IP 192.168.228.17 Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.262271 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.262620 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.363432 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.363534 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.388641 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /readiness Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.389130 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.390698 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/readiness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:33 GMT]] 0xc421f22b20 0 [] true false map[] 0xc420e61800 } Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.390948 42179 prober.go:118] Readiness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.463992 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.464097 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.478129 42179 csi_attacher.go:362] kubernetes.io/csi: attacher.MountDevice successfully requested NodeStageVolume [/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.478433 42179 operation_generator.go:514] MountVolume.MountDevice succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") device mount path "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.478525 42179 operation_generator.go:591] Resizing is not enabled for this volume kubernetes.io/csi/io.netapp.trident.csi^{"name":"pvc-281709ce7e3711e8","protocol":"block"} Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.478548 42179 csi_mounter.go:77] kubernetes.io/csi: mounter.GetPath generated [/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.478553 42179 csi_mounter.go:98] kubernetes.io/csi: Mounter.SetUpAt(/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount) Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.479087 42179 csi_client.go:242] kubernetes.io/csi: calling NodeGetCapabilities rpc Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.482816 42179 csi_mounter.go:161] kubernetes.io/csi: created target path successfully [/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.482891 42179 csi_mounter.go:311] kubernetes.io/csi: volume data file saved successfully [/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/vol_data.json] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.482905 42179 csi_client.go:117] kubernetes.io/csi: calling NodePublishVolume rpc [volid={"name":"pvc-281709ce7e3711e8","protocol":"block"},target_path=/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.502060 42179 csi_mounter.go:217] kubernetes.io/csi: mounter.SetUp successfully requested NodePublish [/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.502126 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.502406 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "pvc-281709ce7e3711e8" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.530794 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.530878 42179 kuberuntime_manager.go:385] No sandbox for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" can be found. Need to start a new one Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.530895 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.530946 42179 kuberuntime_manager.go:579] SyncPod received new pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)", will create a sandbox for it Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.530955 42179 kuberuntime_manager.go:588] Stopping PodSandbox for "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)", will start new one Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.530972 42179 kuberuntime_manager.go:640] Creating sandbox for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.533166 42179 docker_service.go:440] Setting cgroup parent to: "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.699817 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.867720 42179 factory.go:113] Using factory "docker" for container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.868070 42179 docker_sandbox.go:682] Will attempt to re-write config file /var/lib/docker/containers/52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2/resolv.conf with: Jul 02 16:40:33 pike kubelet[42179]: [nameserver 10.96.0.10 search trident.svc.cluster.local svc.cluster.local cluster.local localdomain options ndots:5] Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.868164 42179 plugins.go:412] Calling network plugin cni to set up pod "pod2b_trident" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.868909 42179 cni.go:286] Got netns path /proc/53499/ns/net Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.868944 42179 cni.go:287] Using podns path trident Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.869040 42179 cni.go:256] About to add CNI network cni-loopback (type=loopback) Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.869243 42179 manager.go:997] Added container: "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" (aliases: [k8s_POD_pod2b_trident_2a3383f3-7e38-11e8-a286-005056383598_0 52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2], namespace: "docker") Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.869610 42179 handler.go:325] Added event &{/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2 2018-07-02 20:40:33.534966317 +0000 UTC containerCreation {}} Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.869648 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.872220 42179 cni.go:286] Got netns path /proc/53499/ns/net Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.872241 42179 cni.go:287] Using podns path trident Jul 02 16:40:33 pike kubelet[42179]: I0702 16:40:33.872321 42179 cni.go:256] About to add CNI network k8s-pod-network (type=calico) Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.929 [INFO][53548] calico.go 75: Extracted identifiers EndpointIDs=&utils.WEPIdentifiers{Namespace:"trident", WEPName:"", WorkloadEndpointIdentifiers:names.WorkloadEndpointIdentifiers{Node:"pike", Orchestrator:"k8s", Endpoint:"eth0", Workload:"", Pod:"pod2b", ContainerID:"52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2"}} Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.929 [INFO][53548] calico.go 76: Loaded CNI NetConf NetConfg=types.NetConf{CNIVersion:"0.3.0", Name:"k8s-pod-network", Type:"calico", IPAM:struct { Name string; Type string "json:\"type\""; Subnet string "json:\"subnet\""; AssignIpv4 *string "json:\"assign_ipv4\""; AssignIpv6 *string "json:\"assign_ipv6\""; IPv4Pools []string "json:\"ipv4_pools,omitempty\""; IPv6Pools []string "json:\"ipv6_pools,omitempty\"" }{Name:"", Type:"host-local", Subnet:"usePodCidr", AssignIpv4:(*string)(nil), AssignIpv6:(*string)(nil), IPv4Pools:[]string(nil), IPv6Pools:[]string(nil)}, Args:types.Args{Mesos:types.Mesos{NetworkInfo:types.NetworkInfo{Name:"", Labels:struct { Labels []struct { Key string "json:\"key\""; Value string "json:\"value\"" } "json:\"labels,omitempty\"" }{Labels:[]struct { Key string "json:\"key\""; Value string "json:\"value\"" }(nil)}}}}, MTU:1500, Nodename:"pike", NodenameFileOptional:false, DatastoreType:"kubernetes", EtcdEndpoints:"", LogLevel:"info", Policy:types.Policy{PolicyType:"k8s", K8sAPIRoot:"", K8sAuthToken:"", K8sClientCertificate:"", K8sClientKey:"", K8sCertificateAuthority:""}, Kubernetes:types.Kubernetes{K8sAPIRoot:"", Kubeconfig:"/etc/cni/net.d/calico-kubeconfig", NodeName:""}, EtcdScheme:"", EtcdKeyFile:"", EtcdCertFile:"", EtcdCaCertFile:"", ContainerSettings:types.ContainerSettings{AllowIPForwarding:false}, EtcdAuthority:"", Hostname:""} Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.929 [INFO][53548] utils.go 379: Configured environment: [CNI_COMMAND=ADD CNI_CONTAINERID=52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2 CNI_NETNS=/proc/53499/ns/net CNI_ARGS=IgnoreUnknown=1;IgnoreUnknown=1;K8S_POD_NAMESPACE=trident;K8S_POD_NAME=pod2b;K8S_POD_INFRA_CONTAINER_ID=52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2 CNI_IFNAME=eth0 CNI_PATH=/opt/calico/bin:/opt/cni/bin LANG=en_US.UTF-8 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf KUBELET_SYSTEM_PODS_ARGS=--pod-manifest-path=/etc/kubernetes/manifests --allow-privileged=true KUBELET_NETWORK_ARGS=--network-plugin=cni --cni-conf-dir=/etc/cni/net.d --cni-bin-dir=/opt/cni/bin KUBELET_DNS_ARGS=--cluster-dns=10.96.0.10 --cluster-domain=cluster.local KUBELET_AUTHZ_ARGS=--authorization-mode=Webhook --client-ca-file=/etc/kubernetes/pki/ca.crt KUBELET_CADVISOR_ARGS=--cadvisor-port=0 KUBELET_CERTIFICATE_ARGS=--rotate-certificates=true --cert-dir=/var/lib/kubelet/pki DATASTORE_TYPE=kubernetes KUBECONFIG=/etc/cni/net.d/calico-kubeconfig] Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.976 [INFO][53548] calico.go 167: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {pike-k8s-pod2b-eth0 trident 2a3383f3-7e38-11e8-a286-005056383598 149686 0 2018-07-02 16:40:32 -0400 EDT map[projectcalico.org/namespace:trident projectcalico.org/orchestrator:k8s] map[] [] nil [] } {k8s pike pod2b eth0 [] [] [kns.trident] cali4c32b4784f5 []}} ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-" Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.976 [INFO][53548] k8s.go 60: Extracted identifiers for CmdAddK8s ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" Jul 02 16:40:33 pike kubelet[42179]: Calico CNI fetching podCidr from Kubernetes Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.991 [INFO][53548] k8s.go 83: Fetched podCidr ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" podCidr="192.168.0.0/24" Jul 02 16:40:33 pike kubelet[42179]: Calico CNI passing podCidr to host-local IPAM: 192.168.0.0/24 Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.995 [INFO][53548] k8s.go 660: pod info &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:pod2b,GenerateName:,Namespace:trident,SelfLink:/api/v1/namespaces/trident/pods/pod2b,UID:2a3383f3-7e38-11e8-a286-005056383598,ResourceVersion:149686,Generation:0,CreationTimestamp:2018-07-02 16:40:32 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{data {nil nil nil nil nil nil nil nil nil PersistentVolumeClaimVolumeSource{ClaimName:pvc2,ReadOnly:false,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {default-token-9dv6w {nil nil nil nil nil &SecretVolumeSource{SecretName:default-token-9dv6w,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:latest [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{data false /var/www } {default-token-9dv6w true /var/run/secrets/kubernetes.io/serviceaccount }] nil nil nil /dev/termination-log File Always nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:pike,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc420479530} {node.kubernetes.io/unreachable Exists NoExecute 0xc4204795e0}],HostAliases:[],PriorityClassName:,Priority:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:40:32 -0400 EDT } {Ready False 0001-01-01 00 Jul 02 16:40:33 pike kubelet[42179]: :00:00 +0000 UTC 2018-07-02 16:40:32 -0400 EDT ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:40:32 -0400 EDT }],Message:,Reason:,HostIP:192.168.228.17,PodIP:,StartTime:2018-07-02 16:40:32 -0400 EDT,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:latest }],QOSClass:BestEffort,InitContainerStatuses:[],},} Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.998 [INFO][53548] k8s.go 267: Populated endpoint ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"pike-k8s-pod2b-eth0", GenerateName:"", Namespace:"trident", SelfLink:"", UID:"2a3383f3-7e38-11e8-a286-005056383598", ResourceVersion:"149686", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666160832, loc:(*time.Location)(0x1ec6320)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"trident", "projectcalico.org/orchestrator":"k8s"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"pike", ContainerID:"", Pod:"pod2b", Endpoint:"eth0", IPNetworks:[]string{"192.168.0.54/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.trident"}, InterfaceName:"cali4c32b4784f5", MAC:"", Ports:[]v3.EndpointPort(nil)}} Jul 02 16:40:33 pike kubelet[42179]: Calico CNI using IPs: [192.168.0.54/32] Jul 02 16:40:33 pike kubelet[42179]: 2018-07-02 16:40:33.998 [INFO][53548] network.go 31: Setting the host side veth name to cali4c32b4784f5 ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" Jul 02 16:40:34 pike kubelet[42179]: 2018-07-02 16:40:34.000 [INFO][53548] network.go 326: Disabling IPv4 forwarding ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" Jul 02 16:40:34 pike kubelet[42179]: 2018-07-02 16:40:34.030 [INFO][53548] k8s.go 294: Added Mac, interface name, and active container ID to endpoint ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"pike-k8s-pod2b-eth0", GenerateName:"", Namespace:"trident", SelfLink:"", UID:"2a3383f3-7e38-11e8-a286-005056383598", ResourceVersion:"149686", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666160832, loc:(*time.Location)(0x1ec6320)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"trident", "projectcalico.org/orchestrator":"k8s"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"pike", ContainerID:"52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2", Pod:"pod2b", Endpoint:"eth0", IPNetworks:[]string{"192.168.0.54/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.trident"}, InterfaceName:"cali4c32b4784f5", MAC:"92:cb:b4:db:39:12", Ports:[]v3.EndpointPort(nil)}} Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.047264 42179 config.go:297] Setting pods for source api Jul 02 16:40:34 pike kubelet[42179]: 2018-07-02 16:40:34.052 [INFO][53548] k8s.go 302: Wrote updated endpoint to datastore ContainerID="52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" Namespace="trident" Pod="pod2b" WorkloadEndpoint="pike-k8s-pod2b-eth0" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.053920 42179 kubelet.go:1863] SyncLoop (UPDATE, "api"): "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.066895 42179 kuberuntime_manager.go:654] Created PodSandbox "52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2" for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.071681 42179 kuberuntime_manager.go:673] Determined the ip "192.168.0.54" for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" after sandbox changed Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.071766 42179 kuberuntime_manager.go:724] Creating container &Container{Name:nginx,Image:nginx:latest,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TCP }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{data false /var/www } {default-token-9dv6w true /var/run/secrets/kubernetes.io/serviceaccount }],LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,VolumeDevices:[],} in pod pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598) Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.074103 42179 kuberuntime_image.go:47] Pulling image "nginx:latest" without credentials Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.074136 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx:latest" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.144546 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /liveness Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.144619 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.146092 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/liveness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:34 GMT]] 0xc421ed2d20 0 [] true false map[] 0xc420e61300 } Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.146157 42179 prober.go:118] Liveness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.188092 42179 qos_container_manager_linux.go:317] [ContainerManager]: Updated QoS cgroup configuration Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.241084 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-b950b85ebcfc46dd183ef5f4bd7429bf629b924d6e5966a5b0e9d4c1b7b099b5.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.241363 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-b950b85ebcfc46dd183ef5f4bd7429bf629b924d6e5966a5b0e9d4c1b7b099b5.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.241510 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-b950b85ebcfc46dd183ef5f4bd7429bf629b924d6e5966a5b0e9d4c1b7b099b5.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.241638 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.241761 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.241881 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242007 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/-.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242127 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242246 42179 manager.go:957] ignoring container "/system.slice/-.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242364 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-1b5114bb852c650bff9b089a45465e5d5f84b9e121848e812b890a08a9d2a872.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242484 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-1b5114bb852c650bff9b089a45465e5d5f84b9e121848e812b890a08a9d2a872.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242604 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-1b5114bb852c650bff9b089a45465e5d5f84b9e121848e812b890a08a9d2a872.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242746 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-rpc_pipefs.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242869 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-rpc_pipefs.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.242987 42179 manager.go:957] ignoring container "/system.slice/run-rpc_pipefs.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.243105 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.243259 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.243407 42179 manager.go:957] ignoring container "/system.slice/sys-kernel-debug.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.243548 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/proc-fs-nfsd.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.243737 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/proc-fs-nfsd.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.243893 42179 manager.go:957] ignoring container "/system.slice/proc-fs-nfsd.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244059 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-8f0e06940c1a0eaf129ba7a8e399dac2f41b6fb16b57c0434a9d870e55f7a013.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244183 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-8f0e06940c1a0eaf129ba7a8e399dac2f41b6fb16b57c0434a9d870e55f7a013.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244328 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-8f0e06940c1a0eaf129ba7a8e399dac2f41b6fb16b57c0434a9d870e55f7a013.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244479 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-8eaa19b4\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7enfs-trident\\x2dcsi.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244637 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-8eaa19b4\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7enfs-trident\\x2dcsi.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244817 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-8eaa19b4\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7enfs-trident\\x2dcsi.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.244954 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245075 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245192 42179 manager.go:957] ignoring container "/system.slice/run-user-1000.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245310 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/sys-fs-fuse-connections.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245450 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/sys-fs-fuse-connections.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245569 42179 manager.go:957] ignoring container "/system.slice/sys-fs-fuse-connections.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245687 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-23b9fcad60c832e6c995427b6a15ff6115fc63173e155df233fbae6eec1166be.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245801 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-23b9fcad60c832e6c995427b6a15ff6115fc63173e155df233fbae6eec1166be.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.245930 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-23b9fcad60c832e6c995427b6a15ff6115fc63173e155df233fbae6eec1166be.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.246061 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-bafaa6c3\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-kube\\x2ddns\\x2dtoken\\x2dgbkms.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.246237 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-bafaa6c3\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-kube\\x2ddns\\x2dtoken\\x2dgbkms.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.246415 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-bafaa6c3\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-kube\\x2ddns\\x2dtoken\\x2dgbkms.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.246596 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.246726 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.246880 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-366ecfc8938f2196c91d47118290b8c645348345f7aabdc6eb9ba301b7e9b6c2-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247059 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-vmblock\\x2dfuse.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247179 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-vmblock\\x2dfuse.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247297 42179 manager.go:957] ignoring container "/system.slice/run-vmblock\\x2dfuse.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247447 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-ddf1d3f313126806fc036f4b2448783dfb7e3d80f8da72334123bf2cc4f328e1.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247626 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-ddf1d3f313126806fc036f4b2448783dfb7e3d80f8da72334123bf2cc4f328e1.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247795 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-ddf1d3f313126806fc036f4b2448783dfb7e3d80f8da72334123bf2cc4f328e1.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.247957 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248094 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248215 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248336 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248457 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248578 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-18ea1a7c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248702 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248821 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.248940 42179 manager.go:957] ignoring container "/system.slice/sys-kernel-config.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249057 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-f32b41b009887b275502f2590c2a03f04a1aca9e87f56d650864c78ac73c464a.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249176 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-f32b41b009887b275502f2590c2a03f04a1aca9e87f56d650864c78ac73c464a.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249296 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-f32b41b009887b275502f2590c2a03f04a1aca9e87f56d650864c78ac73c464a.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249419 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-c981c5cbf1d3f260d2ff9f9d0e7ebd39bcd15d334a4615c743fdc49b5016a843.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249539 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-c981c5cbf1d3f260d2ff9f9d0e7ebd39bcd15d334a4615c743fdc49b5016a843.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249659 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-c981c5cbf1d3f260d2ff9f9d0e7ebd39bcd15d334a4615c743fdc49b5016a843.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249789 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-9e6575968d588cdf0f4327f36ac9a103abde98ddb40040b1e340255b46306ea5.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249955 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-9e6575968d588cdf0f4327f36ac9a103abde98ddb40040b1e340255b46306ea5.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249967 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-9e6575968d588cdf0f4327f36ac9a103abde98ddb40040b1e340255b46306ea5.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249973 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-8ed7c19a\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-trident\\x2dcsi\\x2dtoken\\x2d4gzh4.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249979 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-8ed7c19a\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-trident\\x2dcsi\\x2dtoken\\x2d4gzh4.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249986 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-8ed7c19a\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-trident\\x2dcsi\\x2dtoken\\x2d4gzh4.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249991 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-origin-openshift.local.volumes-pods-6bdf507c\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-serving\\x2dcert.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.249997 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-origin-openshift.local.volumes-pods-6bdf507c\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-serving\\x2dcert.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250003 42179 manager.go:957] ignoring container "/system.slice/var-lib-origin-openshift.local.volumes-pods-6bdf507c\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-serving\\x2dcert.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250009 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250013 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250017 42179 manager.go:957] ignoring container "/system.slice/dev-hugepages.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250021 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-a84c8772162c1c84b76d2aef166097277a8905cd98c67f8e0ab7db127c3a3a05.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250025 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-a84c8772162c1c84b76d2aef166097277a8905cd98c67f8e0ab7db127c3a3a05.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250030 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-a84c8772162c1c84b76d2aef166097277a8905cd98c67f8e0ab7db127c3a3a05.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250035 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-fe1240e4\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-calico\\x2dnode\\x2dtoken\\x2dvrxwt.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250040 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-fe1240e4\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-calico\\x2dnode\\x2dtoken\\x2dvrxwt.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250047 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-fe1240e4\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-calico\\x2dnode\\x2dtoken\\x2dvrxwt.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250052 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-0a13356d0619e960e926b2f756856f29f2b29c9d21f7693137e9f996a7eef03c.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250057 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-0a13356d0619e960e926b2f756856f29f2b29c9d21f7693137e9f996a7eef03c.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250062 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-0a13356d0619e960e926b2f756856f29f2b29c9d21f7693137e9f996a7eef03c.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250066 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-9a2770ec2e1fbc84ae9e1cb2b009188e9d0780ece6619af72b4fc1123a25a907.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250071 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-9a2770ec2e1fbc84ae9e1cb2b009188e9d0780ece6619af72b4fc1123a25a907.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250076 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-9a2770ec2e1fbc84ae9e1cb2b009188e9d0780ece6619af72b4fc1123a25a907.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250080 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250084 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250088 42179 manager.go:957] ignoring container "/system.slice/run-docker-netns-default.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250092 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-8eaa19b4\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-trident\\x2dcsi\\x2dtoken\\x2d4gzh4.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250097 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-8eaa19b4\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-trident\\x2dcsi\\x2dtoken\\x2d4gzh4.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250103 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-8eaa19b4\\x2d7e34\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-trident\\x2dcsi\\x2dtoken\\x2d4gzh4.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250109 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-564d228db7fb54f4075f53a8ee37a880195ba85af1d543e4e644847fe94cd3a9.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250113 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-564d228db7fb54f4075f53a8ee37a880195ba85af1d543e4e644847fe94cd3a9.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250118 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-564d228db7fb54f4075f53a8ee37a880195ba85af1d543e4e644847fe94cd3a9.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250123 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-b77d4f8355960ffbc9b3676240cf78a9ef1ac5195fc2a835ed41b0e554b4abbe.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250131 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-b77d4f8355960ffbc9b3676240cf78a9ef1ac5195fc2a835ed41b0e554b4abbe.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250136 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-b77d4f8355960ffbc9b3676240cf78a9ef1ac5195fc2a835ed41b0e554b4abbe.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250141 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-b6d768ffd43273ffc1321e4ece58d107ea5d5111a8d3c9e77344463e473d7e2d-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250146 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-b6d768ffd43273ffc1321e4ece58d107ea5d5111a8d3c9e77344463e473d7e2d-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250151 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-b6d768ffd43273ffc1321e4ece58d107ea5d5111a8d3c9e77344463e473d7e2d-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250156 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-7b54a11572c1c7da4ad457a71c70382c77b418255fda74e1f6b7f3387e7af385.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250160 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-7b54a11572c1c7da4ad457a71c70382c77b418255fda74e1f6b7f3387e7af385.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250165 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-7b54a11572c1c7da4ad457a71c70382c77b418255fda74e1f6b7f3387e7af385.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250170 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-origin-openshift.local.volumes-pods-7446c5fa\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-registry\\x2dtoken\\x2dqdjdl.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250175 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-origin-openshift.local.volumes-pods-7446c5fa\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-registry\\x2dtoken\\x2dqdjdl.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250185 42179 manager.go:957] ignoring container "/system.slice/var-lib-origin-openshift.local.volumes-pods-7446c5fa\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-registry\\x2dtoken\\x2dqdjdl.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250191 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-e55990d6f231ec4409f22309fb58d7ec95004ae35354768ccee9e465c125880d-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250196 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-e55990d6f231ec4409f22309fb58d7ec95004ae35354768ccee9e465c125880d-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250201 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-e55990d6f231ec4409f22309fb58d7ec95004ae35354768ccee9e465c125880d-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250206 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-d038a72d9c856b12903fad635af28e9c9bb2e57e182472e845f6a9e406556885.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250215 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-d038a72d9c856b12903fad635af28e9c9bb2e57e182472e845f6a9e406556885.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250220 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-d038a72d9c856b12903fad635af28e9c9bb2e57e182472e845f6a9e406556885.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250225 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-f0bc7d737878c4ededb7e627c3ba4ef3480dbb6abd608793e0d754d3812159ae-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250230 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-f0bc7d737878c4ededb7e627c3ba4ef3480dbb6abd608793e0d754d3812159ae-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250235 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-f0bc7d737878c4ededb7e627c3ba4ef3480dbb6abd608793e0d754d3812159ae-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250239 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-39a3e139eb579f6d170f362190ee09a1865081755e98b7d788c9850968cdab06-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250244 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-39a3e139eb579f6d170f362190ee09a1865081755e98b7d788c9850968cdab06-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250249 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-39a3e139eb579f6d170f362190ee09a1865081755e98b7d788c9850968cdab06-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250254 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-a17abe3ed6ec.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250258 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-docker-netns-a17abe3ed6ec.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250262 42179 manager.go:957] ignoring container "/system.slice/run-docker-netns-a17abe3ed6ec.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250266 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-7b81eb46dd0fac833eff1ea683b75ba967ebb433b856dbf236825a8fdb8f502b.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250270 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-7b81eb46dd0fac833eff1ea683b75ba967ebb433b856dbf236825a8fdb8f502b.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250275 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-7b81eb46dd0fac833eff1ea683b75ba967ebb433b856dbf236825a8fdb8f502b.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250280 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-user-1000-gvfs.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250283 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-user-1000-gvfs.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250289 42179 manager.go:957] ignoring container "/system.slice/run-user-1000-gvfs.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250320 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-93a44f0375d7db8ef23799ce165a68c8c5fab2aa712fba134808848c7e48d7da.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250328 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-93a44f0375d7db8ef23799ce165a68c8c5fab2aa712fba134808848c7e48d7da.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250347 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-93a44f0375d7db8ef23799ce165a68c8c5fab2aa712fba134808848c7e48d7da.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250353 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-ee3b33397c0d.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250357 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-docker-netns-ee3b33397c0d.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250362 42179 manager.go:957] ignoring container "/system.slice/run-docker-netns-ee3b33397c0d.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250366 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-8a4eb6cd75ba88ccd134ef4342d8d88586760064e2b334115d3d8808d1607715.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250370 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-8a4eb6cd75ba88ccd134ef4342d8d88586760064e2b334115d3d8808d1607715.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250376 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-8a4eb6cd75ba88ccd134ef4342d8d88586760064e2b334115d3d8808d1607715.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250380 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-bb5214c08fbcb1d4c2ab18ad925d05de146ab07e426f9e584f2e3ba44576866a.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250385 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-bb5214c08fbcb1d4c2ab18ad925d05de146ab07e426f9e584f2e3ba44576866a.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250390 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-bb5214c08fbcb1d4c2ab18ad925d05de146ab07e426f9e584f2e3ba44576866a.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250395 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-e6790300b2657e19d1d3c485b69b88ef35a4314e561a2a4854f2440df321813a.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250400 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-e6790300b2657e19d1d3c485b69b88ef35a4314e561a2a4854f2440df321813a.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250405 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-e6790300b2657e19d1d3c485b69b88ef35a4314e561a2a4854f2440df321813a.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250409 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-d716714af9ce066cc00c61812e99a57bb6321a94cd5ea4e5cd804a496d42fef2-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250424 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-d716714af9ce066cc00c61812e99a57bb6321a94cd5ea4e5cd804a496d42fef2-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250429 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-d716714af9ce066cc00c61812e99a57bb6321a94cd5ea4e5cd804a496d42fef2-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250434 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-858767b5ece8ace4630762fef9c2b9e090d6ee8bd9c99fa27a114bb67dd7680c.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250438 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-858767b5ece8ace4630762fef9c2b9e090d6ee8bd9c99fa27a114bb67dd7680c.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250443 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-858767b5ece8ace4630762fef9c2b9e090d6ee8bd9c99fa27a114bb67dd7680c.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250448 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-05e8b02130f444b4ff82853aeada940a02821be0ad667bc4f9585050b9b4da39.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250452 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-05e8b02130f444b4ff82853aeada940a02821be0ad667bc4f9585050b9b4da39.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250457 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-05e8b02130f444b4ff82853aeada940a02821be0ad667bc4f9585050b9b4da39.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250462 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-15428e6a679758a9fcfaae927fe4c120be5e5170dd0e133486eac67638c50455-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250467 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-15428e6a679758a9fcfaae927fe4c120be5e5170dd0e133486eac67638c50455-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250472 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-15428e6a679758a9fcfaae927fe4c120be5e5170dd0e133486eac67638c50455-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250476 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-origin-openshift.local.volumes-pods-6bdf507c\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-webconsole\\x2dtoken\\x2d9c9vc.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250482 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-origin-openshift.local.volumes-pods-6bdf507c\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-webconsole\\x2dtoken\\x2d9c9vc.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250489 42179 manager.go:957] ignoring container "/system.slice/var-lib-origin-openshift.local.volumes-pods-6bdf507c\\x2d74f7\\x2d11e8\\x2d90bc\\x2d005056383598-volumes-kubernetes.io\\x7esecret-webconsole\\x2dtoken\\x2d9c9vc.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250495 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-16d7b0cba4bf7a113e46c1ccc7806138d278f23d8cc560bad09b0687d2b41011-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250499 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-16d7b0cba4bf7a113e46c1ccc7806138d278f23d8cc560bad09b0687d2b41011-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250505 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-16d7b0cba4bf7a113e46c1ccc7806138d278f23d8cc560bad09b0687d2b41011-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250511 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-5edd512a93507d55e044bf1f91b588a436e31744894c70061ca110e708462e38.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250516 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-5edd512a93507d55e044bf1f91b588a436e31744894c70061ca110e708462e38.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250521 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-5edd512a93507d55e044bf1f91b588a436e31744894c70061ca110e708462e38.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250526 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-620482ceb4b288dcf7079ed286c5e45e94af11172d4da0fbb9ae7b1c8ec439f0.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250530 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-620482ceb4b288dcf7079ed286c5e45e94af11172d4da0fbb9ae7b1c8ec439f0.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250537 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-620482ceb4b288dcf7079ed286c5e45e94af11172d4da0fbb9ae7b1c8ec439f0.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250542 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250546 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250550 42179 manager.go:957] ignoring container "/system.slice/dev-mqueue.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250553 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-fdf93005011de0e9a6e99298bacb4068da25a044851e342c364e9a1807acc085-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250558 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-fdf93005011de0e9a6e99298bacb4068da25a044851e342c364e9a1807acc085-shm.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250563 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-fdf93005011de0e9a6e99298bacb4068da25a044851e342c364e9a1807acc085-shm.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250568 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-c08cc68d869222039a14a1bafd6138fce75bc25ef627fa96bb15c404d28d4654.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250572 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-c08cc68d869222039a14a1bafd6138fce75bc25ef627fa96bb15c404d28d4654.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250577 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-c08cc68d869222039a14a1bafd6138fce75bc25ef627fa96bb15c404d28d4654.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250582 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-baeff5dc\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dq4qcr.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250588 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-baeff5dc\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dq4qcr.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250594 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-baeff5dc\\x2d773a\\x2d11e8\\x2da95f\\x2d005056383598-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dq4qcr.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250600 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-08d3087f02fa.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250603 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-docker-netns-08d3087f02fa.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250608 42179 manager.go:957] ignoring container "/system.slice/run-docker-netns-08d3087f02fa.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250611 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-34d5205d1c7335c3cd1aa050d545a0ffba499ec20e3f9ed921facaed28979e95.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250616 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-34d5205d1c7335c3cd1aa050d545a0ffba499ec20e3f9ed921facaed28979e95.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250621 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-34d5205d1c7335c3cd1aa050d545a0ffba499ec20e3f9ed921facaed28979e95.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250625 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-b1b267d9f4dbbeed6fa978106eac958f77aac402567e8570e12681090156d9cb.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250630 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-b1b267d9f4dbbeed6fa978106eac958f77aac402567e8570e12681090156d9cb.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250635 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-b1b267d9f4dbbeed6fa978106eac958f77aac402567e8570e12681090156d9cb.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250639 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-origin-openshift.local.volumes.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250643 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-origin-openshift.local.volumes.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250648 42179 manager.go:957] ignoring container "/system.slice/var-lib-origin-openshift.local.volumes.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250652 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-plugins.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250658 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-plugins.mount", but ignoring. Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.250662 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-plugins.mount" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.526940 42179 generic.go:147] GenericPLEG: 2a3383f3-7e38-11e8-a286-005056383598/52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2: non-existent -> running Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.529484 42179 kuberuntime_manager.go:852] getSandboxIDByPodUID got sandbox IDs ["52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2"] for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.534262 42179 generic.go:380] PLEG: Write status for pod2b/trident: &container.PodStatus{ID:"2a3383f3-7e38-11e8-a286-005056383598", Name:"pod2b", Namespace:"trident", IP:"192.168.0.54", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*v1alpha2.PodSandboxStatus{(*v1alpha2.PodSandboxStatus)(0xc4214522d0)}} (err: ) Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.534511 42179 kubelet.go:1901] SyncLoop (PLEG): "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)", event: &pleg.PodLifecycleEvent{ID:"2a3383f3-7e38-11e8-a286-005056383598", Type:"ContainerStarted", Data:"52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2"} Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.614230 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /metrics Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.614332 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.619417 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; version=0.0.4] Date:[Mon, 02 Jul 2018 20:40:34 GMT]] 0xc421812be0 -1 [] true true map[] 0xc422b62500 } Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.619482 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):sidecar" succeeded Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.681734 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.697954 42179 kube_docker_client.go:348] Stop pulling image "nginx:latest": "Status: Image is up to date for nginx:latest" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.699479 42179 kuberuntime_container.go:100] Generating ref for container nginx: &v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:"spec.containers{nginx}"} Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.699569 42179 kubelet_pods.go:172] container: trident/pod2b/nginx podIP: "192.168.0.54" creating hosts mount: true Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.699595 42179 csi_mounter.go:77] kubernetes.io/csi: mounter.GetPath generated [/var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.699809 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "nginx:latest" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.702414 42179 docker_service.go:440] Setting cgroup parent to: "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.810899 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Created' Created container Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.952441 42179 factory.go:113] Using factory "docker" for container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c" Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.953747 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2b", UID:"2a3383f3-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149684", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Started' Started container Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.955360 42179 manager.go:997] Added container: "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c" (aliases: [k8s_nginx_pod2b_trident_2a3383f3-7e38-11e8-a286-005056383598_0 ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c], namespace: "docker") Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.955656 42179 handler.go:325] Added event &{/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c 2018-07-02 20:40:34.706298597 +0000 UTC containerCreation {}} Jul 02 16:40:34 pike kubelet[42179]: I0702 16:40:34.955877 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod2a3383f3-7e38-11e8-a286-005056383598/ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c" Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.547868 42179 generic.go:147] GenericPLEG: 2a3383f3-7e38-11e8-a286-005056383598/ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c: non-existent -> running Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.549520 42179 kuberuntime_manager.go:852] getSandboxIDByPodUID got sandbox IDs ["52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2"] for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.557458 42179 generic.go:380] PLEG: Write status for pod2b/trident: &container.PodStatus{ID:"2a3383f3-7e38-11e8-a286-005056383598", Name:"pod2b", Namespace:"trident", IP:"192.168.0.54", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc421dd8c40)}, SandboxStatuses:[]*v1alpha2.PodSandboxStatus{(*v1alpha2.PodSandboxStatus)(0xc4219f5ef0)}} (err: ) Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.557557 42179 kubelet.go:1901] SyncLoop (PLEG): "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)", event: &pleg.PodLifecycleEvent{ID:"2a3383f3-7e38-11e8-a286-005056383598", Type:"ContainerStarted", Data:"ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c"} Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.557592 42179 kubelet_pods.go:1366] Generating status for "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.557794 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.564874 42179 config.go:297] Setting pods for source api Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.564896 42179 status_manager.go:481] Status for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:35 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.54 StartTime:2018-07-02 16:40:32 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:40:34 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:nginx:latest ImageID:docker-pullable://nginx@sha256:62a095e5da5f977b9f830adaf64d604c614024bf239d21068e4ca826d0d629a4 ContainerID:docker://ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c}] QOSClass:BestEffort}) Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.565600 42179 kubelet.go:1869] SyncLoop (RECONCILE, "api"): "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.641984 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "2a3383f3-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.642064 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "2a3383f3-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.675570 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.675803 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 2a3383f3-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.679194 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.679449 42179 atomic_writer.go:156] pod trident/pod2b volume default-token-9dv6w: no update required for target directory /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.679472 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.699912 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.858240 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:35 pike kubelet[42179]: I0702 16:40:35.858387 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.570095 42179 kubelet_pods.go:1366] Generating status for "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.570320 42179 status_manager.go:340] Ignoring same status for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:35 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:32 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.54 StartTime:2018-07-02 16:40:32 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:40:34 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:nginx:latest ImageID:docker-pullable://nginx@sha256:62a095e5da5f977b9f830adaf64d604c614024bf239d21068e4ca826d0d629a4 ContainerID:docker://ef61c1edcf671161fa81730d58ca5a9c4ab1bf1ac6c9fe502c5df5e19264e75c}] QOSClass:BestEffort} Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.570596 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.654168 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "2a3383f3-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.654242 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "2a3383f3-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.680992 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.681191 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 2a3383f3-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.682899 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.683076 42179 atomic_writer.go:156] pod trident/pod2b volume default-token-9dv6w: no update required for target directory /var/lib/kubelet/pods/2a3383f3-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.683152 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2a3383f3-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2b" (UID: "2a3383f3-7e38-11e8-a286-005056383598") Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.871697 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:36 pike kubelet[42179]: I0702 16:40:36.871938 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "pod2b_trident(2a3383f3-7e38-11e8-a286-005056383598)" Jul 02 16:40:37 pike kubelet[42179]: I0702 16:40:37.029887 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz Jul 02 16:40:37 pike kubelet[42179]: I0702 16:40:37.030476 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:37 pike kubelet[42179]: I0702 16:40:37.032600 42179 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420ff89e0 2 [] true false map[] 0xc420a9a700 } Jul 02 16:40:37 pike kubelet[42179]: I0702 16:40:37.032978 42179 prober.go:118] Liveness probe for "kube-controller-manager-pike_kube-system(fac56382eca21a00fc5c8a039563b311):kube-controller-manager" succeeded Jul 02 16:40:37 pike kubelet[42179]: I0702 16:40:37.700118 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.207296 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/dnsmasq Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.207399 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.208281 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[49] Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:39 GMT]] 0xc420d4a320 49 [] true false map[] 0xc420a9b900 } Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.208359 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):dnsmasq" succeeded Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.684777 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.700321 42179 kubelet.go:1920] SyncLoop (SYNC): 1 pods; kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e) Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.700438 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.700683 42179 kubelet_pods.go:1366] Generating status for "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e)" Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.700920 42179 status_manager.go:340] Ignoring same status for pod "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-23 19:09:51 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-29 12:55:56 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-06-23 19:09:51 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.228.17 StartTime:2018-06-23 19:09:51 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-06-29 12:55:27 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:2,Signal:0,Reason:Error,Message:,StartedAt:2018-06-29 12:46:22 -0400 EDT,FinishedAt:2018-06-29 12:50:27 -0400 EDT,ContainerID:docker://eb06c4d27c019a8773ec4c24a4c89c17b9022231482f66deded9070d8cd04231,}} Ready:true RestartCount:4 Image:k8s.gcr.io/kube-scheduler-amd64:v1.10.5 ImageID:docker-pullable://k8s.gcr.io/kube-scheduler-amd64@sha256:ccd7da1c35fefdb8077f80baf0724b861b94b3fc182ae0b5e0b7644257a0dd41 ContainerID:docker://319e203936a07b7c6d39a5a76e1e5a860a5d4c6e67f991898ae92f2c01cb73ec}] QOSClass:Burstable} Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.701160 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e)" Jul 02 16:40:39 pike kubelet[42179]: I0702 16:40:39.775681 42179 desired_state_of_world_populator.go:302] Added volume "kubeconfig" (volSpec="kubeconfig") for pod "555ca8f25ef54ae21d81d1336633c27e" to desired state. Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.002336 42179 volume_manager.go:380] All volumes are attached and mounted for pod "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e)" Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.002540 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:16d7b0cba4bf7a113e46c1ccc7806138d278f23d8cc560bad09b0687d2b41011 Attempt:2 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e)" Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.180480 42179 eviction_manager.go:229] eviction manager: synchronize housekeeping Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240681 42179 helpers.go:840] eviction manager: observations: signal=allocatableMemory.available, available: 5865972Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:40.240543108 -0400 EDT m=+1028.774232482 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240738 42179 helpers.go:840] eviction manager: observations: signal=nodefs.available, available: 10751868Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:40.200700485 -0400 EDT m=+1028.734389855 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240750 42179 helpers.go:840] eviction manager: observations: signal=nodefs.inodesFree, available: 1687803, capacity: 2240Ki, time: 2018-07-02 16:40:40.200700485 -0400 EDT m=+1028.734389855 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240756 42179 helpers.go:840] eviction manager: observations: signal=imagefs.available, available: 10751868Ki, capacity: 35992192Ki, time: 2018-07-02 16:40:40.200700485 -0400 EDT m=+1028.734389855 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240762 42179 helpers.go:840] eviction manager: observations: signal=imagefs.inodesFree, available: 1687803, capacity: 2240Ki, time: 2018-07-02 16:40:40.200700485 -0400 EDT m=+1028.734389855 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240768 42179 helpers.go:840] eviction manager: observations: signal=pid.available, available: 129756, capacity: 128Ki, time: 2018-07-02 16:40:40.236114519 -0400 EDT m=+1028.769803873 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240774 42179 helpers.go:840] eviction manager: observations: signal=memory.available, available: 1537984Ki, capacity: 6531384Ki, time: 2018-07-02 16:40:40.200700485 -0400 EDT m=+1028.734389855 Jul 02 16:40:40 pike kubelet[42179]: I0702 16:40:40.240830 42179 eviction_manager.go:336] eviction manager: no resources are starved Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.044862 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /healthcheck/kubedns Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.044911 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.046265 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[application/json] Date:[Mon, 02 Jul 2018 20:40:41 GMT] Content-Length:[51]] 0xc421813f40 51 [] true false map[] 0xc420646300 } Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.046404 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.162893 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:etcd-pike,GenerateName:,Namespace:kube-system,SelfLink:/api/v1/namespaces/kube-system/pods/etcd-pike,UID:68f9c58845a7c8f7ac9d6261d498ab6c,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{component: etcd,tier: control-plane,},Annotations:map[string]string{kubernetes.io/config.hash: 68f9c58845a7c8f7ac9d6261d498ab6c,kubernetes.io/config.seen: 2018-07-02T16:23:31.6928551-04:00,kubernetes.io/config.source: file,scheduler.alpha.kubernetes.io/critical-pod: ,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-data {HostPathVolumeSource{Path:/var/lib/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etcd-certs {&HostPathVolumeSource{Path:/etc/kubernetes/pki/etcd,Type:*DirectoryOrCreate,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/h Jul 02 16:40:41 pike kubelet[42179]: ealthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:pike,HostNetwork:true,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{ Exists NoExecute }],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd k8s.gcr.io/etcd-amd64:3.1.12 [etcd --advertise-client-urls=https://127.0.0.1:2379 --peer-client-cert-auth=true --cert-file=/etc/kubernetes/pki/etcd/server.crt --key-file=/etc/kubernetes/pki/etcd/server.key --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --listen-client-urls=https://127.0.0.1:2379 --data-dir=/var/lib/etcd --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --client-cert-auth=true] [] [] [] [] {map[] map[]} [{etcd-data false /var/lib/etcd } {etcd-certs false /etc/kubernetes/pki/etcd }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds: Jul 02 16:40:41 pike kubelet[42179]: 10,SuccessThreshold:1,FailureThreshold:8,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [/bin/sh -ec ETCDCTL_API=3 etcdctl --endpoints=127.0.0.1:2379 --cacert=/etc/kubernetes/pki/etcd/ca.crt --cert=/etc/kubernetes/pki/etcd/healthcheck-client.crt --key=/etc/kubernetes/pki/etcd/healthcheck-client.key get foo] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.220358 42179 exec.go:38] Exec probe response: "2018-07-02 20:40:41.196263 I | warning: ignoring ServerName for user-provided CA for backwards compatibility is deprecated\n" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.220386 42179 prober.go:118] Liveness probe for "etcd-pike_kube-system(68f9c58845a7c8f7ac9d6261d498ab6c):etcd" succeeded Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.251977 42179 prober.go:165] HTTP-Probe Host: https://192.168.228.17, Port: 6443, Path: /healthz Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.252013 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.258564 42179 http.go:96] Probe succeeded for https://192.168.228.17:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Mon, 02 Jul 2018 20:40:41 GMT]] 0xc421feec20 2 [] false false map[] 0xc421d6e300 0xc42119dc30} Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.258629 42179 prober.go:118] Liveness probe for "kube-apiserver-pike_kube-system(89228f8427ffc79800b43469845b379e):kube-apiserver" succeeded Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.317754 42179 config.go:297] Setting pods for source api Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.318218 42179 config.go:405] Receiving a new pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.318324 42179 kubelet.go:1856] SyncLoop (ADD, "api"): "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.331420 42179 kubelet_pods.go:1366] Generating status for "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.332083 42179 qos_container_manager_linux.go:317] [ContainerManager]: Updated QoS cgroup configuration Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333416 42179 factory.go:117] Factory "docker" was unable to handle container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333434 42179 factory.go:106] Error trying to work out if we can handle /kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598: /kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598 not handled by systemd handler Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333441 42179 factory.go:117] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333447 42179 factory.go:113] Using factory "raw" for container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333545 42179 manager.go:997] Added container: "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" (aliases: [], namespace: "") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333599 42179 handler.go:325] Added event &{/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598 2018-07-02 16:40:41.329395989 -0400 EDT containerCreation {}} Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.333625 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.339441 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.347921 42179 config.go:297] Setting pods for source api Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.348301 42179 status_manager.go:481] Status for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" updated successfully: (1, {Phase:Pending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP: StartTime:2018-07-02 16:40:41 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:nginx:latest ImageID: ContainerID:}] QOSClass:BestEffort}) Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.348507 42179 kubelet.go:1869] SyncLoop (RECONCILE, "api"): "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.389014 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "2f638a5c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.389081 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "2f638a5c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410275 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410390 42179 csi_plugin.go:108] kubernetes.io/csi: setting up mounter for [volume={"name":"pvc-281709ce7e3711e8","protocol":"block"},driver=io.netapp.trident.csi] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410461 42179 reconciler.go:252] operationExecutor.MountVolume started for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410515 42179 reconciler.go:207] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410613 42179 operation_generator.go:486] MountVolume.WaitForAttach entering for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") DevicePath "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410639 42179 csi_attacher.go:131] kubernetes.io/csi: probing for updates from CSI driver for [attachment.ID=csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.410661 42179 csi_attacher.go:140] kubernetes.io/csi: probing VolumeAttachment [id=csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.413357 42179 operation_generator.go:495] MountVolume.WaitForAttach succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") DevicePath "csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.413528 42179 csi_attacher.go:244] kubernetes.io/csi: attacher.GetDeviceMountPath(&{nil &PersistentVolume{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:pvc-281709ce7e3711e8,GenerateName:,Namespace:,SelfLink:/api/v1/persistentvolumes/pvc-281709ce7e3711e8,UID:2c19eed1-7e37-11e8-a286-005056383598,ResourceVersion:149392,Generation:0,CreationTimestamp:2018-07-02 16:33:26 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{pv.kubernetes.io/provisioned-by: io.netapp.trident.csi,},OwnerReferences:[],Finalizers:[kubernetes.io/pv-protection external-attacher/io-netapp-trident-csi],ClusterName:,Initializers:nil,},Spec:PersistentVolumeSpec{Capacity:ResourceList{storage: {{1073741824 0} {} 1Gi BinarySI},},PersistentVolumeSource:PersistentVolumeSource{GCEPersistentDisk:nil,AWSElasticBlockStore:nil,HostPath:nil,Glusterfs:nil,NFS:nil,RBD:nil,ISCSI:nil,Cinder:nil,CephFS:nil,FC:nil,Flocker:nil,FlexVolume:nil,AzureFile:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Local:nil,StorageOS:nil,CSI:&CSIPersistentVolumeSource{Driver:io.netapp.trident.csi,VolumeHandle:{"name":"pvc-281709ce7e3711e8","protocol":"block"},ReadOnly:false,FSType:,VolumeAttributes:map[string]string{backend: solidfire_10.250.118.144,internalName: pvc-281709ce7e3711e8,name: pvc-281709ce7e3711e8,protocol: block,storage.kubernetes.io/csiProvisionerIdentity: 1530562488326-8081-io.netapp.trident.csi,},ControllerPublishSecretRef:nil,NodeStageSecretRef:nil,NodePublishSecretRef:nil,},},AccessModes:[ReadWriteMany],ClaimRef:&ObjectReference{Kind:PersistentVolumeClaim,Namespace:trident,Name:pvc2,UID:281709ce-7e37-11e8-a286-005056383598,APIVersion:v1,ResourceVersion:149109,FieldPath:,},PersistentVolumeReclaimPolicy:Delete,StorageClassName:ssd-sf,MountOptions:[],VolumeMode:nil,NodeAffinity:nil,},Status:PersistentVolumeStatus{Phase:Bound,Message:,Reason:,},} false}) Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.413603 42179 csi_attacher.go:250] attacher.GetDeviceMountPath succeeded, deviceMountPath: /var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.413613 42179 csi_attacher.go:255] kubernetes.io/csi: attacher.MountDevice(csi-60bd46039fe9df350921586e47148d64079001f91ecd4e46dbd05a425c873b8c, /var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount) Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.413644 42179 csi_client.go:242] kubernetes.io/csi: calling NodeGetCapabilities rpc Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.416965 42179 csi_attacher.go:326] kubernetes.io/csi: created target path successfully [/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.417010 42179 csi_client.go:186] kubernetes.io/csi: calling NodeStageVolume rpc [volid={"name":"pvc-281709ce7e3711e8","protocol":"block"},staging_target_path=/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.444906 42179 csi_attacher.go:362] kubernetes.io/csi: attacher.MountDevice successfully requested NodeStageVolume [/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.444986 42179 operation_generator.go:514] MountVolume.MountDevice succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") device mount path "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-281709ce7e3711e8/globalmount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.445006 42179 operation_generator.go:591] Resizing is not enabled for this volume kubernetes.io/csi/io.netapp.trident.csi^{"name":"pvc-281709ce7e3711e8","protocol":"block"} Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.445036 42179 csi_mounter.go:77] kubernetes.io/csi: mounter.GetPath generated [/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.445046 42179 csi_mounter.go:98] kubernetes.io/csi: Mounter.SetUpAt(/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount) Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.445092 42179 csi_client.go:242] kubernetes.io/csi: calling NodeGetCapabilities rpc Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.448221 42179 csi_mounter.go:161] kubernetes.io/csi: created target path successfully [/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.448311 42179 csi_mounter.go:311] kubernetes.io/csi: volume data file saved successfully [/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/vol_data.json] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.448323 42179 csi_client.go:117] kubernetes.io/csi: calling NodePublishVolume rpc [volid={"name":"pvc-281709ce7e3711e8","protocol":"block"},target_path=/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.469212 42179 csi_mounter.go:217] kubernetes.io/csi: mounter.SetUp successfully requested NodePublish [/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.469442 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "pvc-281709ce7e3711e8" (UniqueName: "kubernetes.io/csi/io.netapp.trident.csi^{\"name\":\"pvc-281709ce7e3711e8\",\"protocol\":\"block\"}") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.469593 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "pvc-281709ce7e3711e8" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.499663 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 8081, Path: /readiness Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.499714 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.500640 42179 http.go:96] Probe succeeded for http://192.168.0.43:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:41 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc42200c600 3 [] true false map[] 0xc421203000 } Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.500715 42179 prober.go:118] Readiness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):kubedns" succeeded Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.511443 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.511567 42179 reconciler.go:252] operationExecutor.MountVolume started for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.511633 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 2f638a5c-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.511779 42179 empty_dir.go:256] pod 2f638a5c-7e38-11e8-a286-005056383598: mounting tmpfs for volume wrapped_default-token-9dv6w Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.511800 42179 mount_linux.go:143] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w]) Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516231 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516255 42179 factory.go:106] Error trying to work out if we can handle /system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope: /system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope not handled by systemd handler Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516262 42179 factory.go:117] Factory "systemd" was unable to handle container "/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516268 42179 factory.go:113] Using factory "raw" for container "/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516514 42179 manager.go:997] Added container: "/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope" (aliases: [], namespace: "") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516638 42179 handler.go:325] Added event &{/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope 2018-07-02 16:40:41.513387668 -0400 EDT containerCreation {}} Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.516673 42179 container.go:448] Start housekeeping for container "/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.521980 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-259f31a8329da6fa8547e6fbcb53862db4b1992bde332f4ef0c253fc6e23c35c.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.522002 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-259f31a8329da6fa8547e6fbcb53862db4b1992bde332f4ef0c253fc6e23c35c.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.522011 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-259f31a8329da6fa8547e6fbcb53862db4b1992bde332f4ef0c253fc6e23c35c.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.523427 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-c14eac94899f.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.523467 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/run-docker-netns-c14eac94899f.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.523474 42179 manager.go:957] ignoring container "/system.slice/run-docker-netns-c14eac94899f.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524210 42179 prober.go:165] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524239 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524248 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524343 42179 atomic_writer.go:176] pod trident/pod2c volume default-token-9dv6w: performed write of new data to ts data directory: /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w/..2018_07_02_20_40_41.618111285 Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524401 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524840 42179 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Mon, 02 Jul 2018 20:40:41 GMT] Content-Length:[2]] 0xc421ccae20 2 [] true false map[] 0xc421203200 } Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524871 42179 prober.go:118] Liveness probe for "kube-scheduler-pike_kube-system(555ca8f25ef54ae21d81d1336633c27e):kube-scheduler" succeeded Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.524889 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-9dv6w" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.529456 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-aufs-mnt-9b72b26af4810b9e923b3041ce6c0ee5f39dde0c5e03d6e9a120aeb7ccb0e664.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.529502 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-aufs-mnt-9b72b26af4810b9e923b3041ce6c0ee5f39dde0c5e03d6e9a120aeb7ccb0e664.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.529513 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-aufs-mnt-9b72b26af4810b9e923b3041ce6c0ee5f39dde0c5e03d6e9a120aeb7ccb0e664.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.529980 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-2f638a5c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.529993 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-2f638a5c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.530000 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-2f638a5c\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.531095 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-2a3383f3\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.531131 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-2a3383f3\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.531140 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-2a3383f3\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2d9dv6w.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.531818 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2-shm.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.531858 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2-shm.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.531869 42179 manager.go:957] ignoring container "/system.slice/var-lib-docker-containers-52f571f7f8069d00f259ed91da4a7b4c24371e17bf882c3f3180c25a2c8eb9a2-shm.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.532309 42179 factory.go:117] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-2a3383f3\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.532352 42179 factory.go:110] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-2a3383f3\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount", but ignoring. Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.532363 42179 manager.go:957] ignoring container "/system.slice/var-lib-kubelet-pods-2a3383f3\\x2d7e38\\x2d11e8\\x2da286\\x2d005056383598-volumes-kubernetes.io\\x7ecsi-pvc\\x2d281709ce7e3711e8-mount.mount" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.533098 42179 manager.go:1054] Destroyed container: "/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope" (aliases: [], namespace: "") Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.533139 42179 handler.go:325] Added event &{/system.slice/run-r6a31ffbf77384be58c24c6f1170f9fea.scope 2018-07-02 16:40:41.533131079 -0400 EDT m=+1030.066820445 containerDeletion {}} Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.639925 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.639988 42179 kuberuntime_manager.go:385] No sandbox for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" can be found. Need to start a new one Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.640006 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.640051 42179 kuberuntime_manager.go:579] SyncPod received new pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)", will create a sandbox for it Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.640060 42179 kuberuntime_manager.go:588] Stopping PodSandbox for "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)", will start new one Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.640076 42179 kuberuntime_manager.go:640] Creating sandbox for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.641757 42179 docker_service.go:440] Setting cgroup parent to: "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" Jul 02 16:40:41 pike kubelet[42179]: I0702 16:40:41.699947 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.032021 42179 factory.go:113] Using factory "docker" for container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.032262 42179 docker_sandbox.go:682] Will attempt to re-write config file /var/lib/docker/containers/d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1/resolv.conf with: Jul 02 16:40:42 pike kubelet[42179]: [nameserver 10.96.0.10 search trident.svc.cluster.local svc.cluster.local cluster.local localdomain options ndots:5] Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.032351 42179 plugins.go:412] Calling network plugin cni to set up pod "pod2c_trident" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.033338 42179 manager.go:997] Added container: "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" (aliases: [k8s_POD_pod2c_trident_2f638a5c-7e38-11e8-a286-005056383598_0 d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1], namespace: "docker") Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.033563 42179 handler.go:325] Added event &{/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1 2018-07-02 20:40:41.643991674 +0000 UTC containerCreation {}} Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.033628 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.036206 42179 cni.go:286] Got netns path /proc/53863/ns/net Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.036235 42179 cni.go:287] Using podns path trident Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.036336 42179 cni.go:256] About to add CNI network cni-loopback (type=loopback) Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.038633 42179 cni.go:286] Got netns path /proc/53863/ns/net Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.038668 42179 cni.go:287] Using podns path trident Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.038795 42179 cni.go:256] About to add CNI network k8s-pod-network (type=calico) Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.094 [INFO][53916] calico.go 75: Extracted identifiers EndpointIDs=&utils.WEPIdentifiers{Namespace:"trident", WEPName:"", WorkloadEndpointIdentifiers:names.WorkloadEndpointIdentifiers{Node:"pike", Orchestrator:"k8s", Endpoint:"eth0", Workload:"", Pod:"pod2c", ContainerID:"d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1"}} Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.094 [INFO][53916] calico.go 76: Loaded CNI NetConf NetConfg=types.NetConf{CNIVersion:"0.3.0", Name:"k8s-pod-network", Type:"calico", IPAM:struct { Name string; Type string "json:\"type\""; Subnet string "json:\"subnet\""; AssignIpv4 *string "json:\"assign_ipv4\""; AssignIpv6 *string "json:\"assign_ipv6\""; IPv4Pools []string "json:\"ipv4_pools,omitempty\""; IPv6Pools []string "json:\"ipv6_pools,omitempty\"" }{Name:"", Type:"host-local", Subnet:"usePodCidr", AssignIpv4:(*string)(nil), AssignIpv6:(*string)(nil), IPv4Pools:[]string(nil), IPv6Pools:[]string(nil)}, Args:types.Args{Mesos:types.Mesos{NetworkInfo:types.NetworkInfo{Name:"", Labels:struct { Labels []struct { Key string "json:\"key\""; Value string "json:\"value\"" } "json:\"labels,omitempty\"" }{Labels:[]struct { Key string "json:\"key\""; Value string "json:\"value\"" }(nil)}}}}, MTU:1500, Nodename:"pike", NodenameFileOptional:false, DatastoreType:"kubernetes", EtcdEndpoints:"", LogLevel:"info", Policy:types.Policy{PolicyType:"k8s", K8sAPIRoot:"", K8sAuthToken:"", K8sClientCertificate:"", K8sClientKey:"", K8sCertificateAuthority:""}, Kubernetes:types.Kubernetes{K8sAPIRoot:"", Kubeconfig:"/etc/cni/net.d/calico-kubeconfig", NodeName:""}, EtcdScheme:"", EtcdKeyFile:"", EtcdCertFile:"", EtcdCaCertFile:"", ContainerSettings:types.ContainerSettings{AllowIPForwarding:false}, EtcdAuthority:"", Hostname:""} Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.094 [INFO][53916] utils.go 379: Configured environment: [CNI_COMMAND=ADD CNI_CONTAINERID=d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1 CNI_NETNS=/proc/53863/ns/net CNI_ARGS=IgnoreUnknown=1;IgnoreUnknown=1;K8S_POD_NAMESPACE=trident;K8S_POD_NAME=pod2c;K8S_POD_INFRA_CONTAINER_ID=d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1 CNI_IFNAME=eth0 CNI_PATH=/opt/calico/bin:/opt/cni/bin LANG=en_US.UTF-8 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf KUBELET_SYSTEM_PODS_ARGS=--pod-manifest-path=/etc/kubernetes/manifests --allow-privileged=true KUBELET_NETWORK_ARGS=--network-plugin=cni --cni-conf-dir=/etc/cni/net.d --cni-bin-dir=/opt/cni/bin KUBELET_DNS_ARGS=--cluster-dns=10.96.0.10 --cluster-domain=cluster.local KUBELET_AUTHZ_ARGS=--authorization-mode=Webhook --client-ca-file=/etc/kubernetes/pki/ca.crt KUBELET_CADVISOR_ARGS=--cadvisor-port=0 KUBELET_CERTIFICATE_ARGS=--rotate-certificates=true --cert-dir=/var/lib/kubelet/pki DATASTORE_TYPE=kubernetes KUBECONFIG=/etc/cni/net.d/calico-kubeconfig] Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.138 [INFO][53916] calico.go 167: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {pike-k8s-pod2c-eth0 trident 2f638a5c-7e38-11e8-a286-005056383598 149707 0 2018-07-02 16:40:41 -0400 EDT map[projectcalico.org/namespace:trident projectcalico.org/orchestrator:k8s] map[] [] nil [] } {k8s pike pod2c eth0 [] [] [kns.trident] cali804c3f63559 []}} ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-" Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.138 [INFO][53916] k8s.go 60: Extracted identifiers for CmdAddK8s ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" Jul 02 16:40:42 pike kubelet[42179]: Calico CNI fetching podCidr from Kubernetes Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.152 [INFO][53916] k8s.go 83: Fetched podCidr ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" podCidr="192.168.0.0/24" Jul 02 16:40:42 pike kubelet[42179]: Calico CNI passing podCidr to host-local IPAM: 192.168.0.0/24 Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.155 [INFO][53916] k8s.go 660: pod info &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:pod2c,GenerateName:,Namespace:trident,SelfLink:/api/v1/namespaces/trident/pods/pod2c,UID:2f638a5c-7e38-11e8-a286-005056383598,ResourceVersion:149707,Generation:0,CreationTimestamp:2018-07-02 16:40:41 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{data {nil nil nil nil nil nil nil nil nil PersistentVolumeClaimVolumeSource{ClaimName:pvc2,ReadOnly:false,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {default-token-9dv6w {nil nil nil nil nil &SecretVolumeSource{SecretName:default-token-9dv6w,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:latest [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{data false /var/www } {default-token-9dv6w true /var/run/secrets/kubernetes.io/serviceaccount }] nil nil nil /dev/termination-log File Always nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:pike,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc42045ab40} {node.kubernetes.io/unreachable Exists NoExecute 0xc42045ab80}],HostAliases:[],PriorityClassName:,Priority:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:40:41 -0400 EDT } {Ready False 0001-01-01 00 Jul 02 16:40:42 pike kubelet[42179]: :00:00 +0000 UTC 2018-07-02 16:40:41 -0400 EDT ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:40:41 -0400 EDT }],Message:,Reason:,HostIP:192.168.228.17,PodIP:,StartTime:2018-07-02 16:40:41 -0400 EDT,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:latest }],QOSClass:BestEffort,InitContainerStatuses:[],},} Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.157 [INFO][53916] k8s.go 267: Populated endpoint ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"pike-k8s-pod2c-eth0", GenerateName:"", Namespace:"trident", SelfLink:"", UID:"2f638a5c-7e38-11e8-a286-005056383598", ResourceVersion:"149707", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666160841, loc:(*time.Location)(0x1ec6320)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"trident", "projectcalico.org/orchestrator":"k8s"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"pike", ContainerID:"", Pod:"pod2c", Endpoint:"eth0", IPNetworks:[]string{"192.168.0.55/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.trident"}, InterfaceName:"cali804c3f63559", MAC:"", Ports:[]v3.EndpointPort(nil)}} Jul 02 16:40:42 pike kubelet[42179]: Calico CNI using IPs: [192.168.0.55/32] Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.157 [INFO][53916] network.go 31: Setting the host side veth name to cali804c3f63559 ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.159 [INFO][53916] network.go 326: Disabling IPv4 forwarding ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.185 [INFO][53916] k8s.go 294: Added Mac, interface name, and active container ID to endpoint ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"pike-k8s-pod2c-eth0", GenerateName:"", Namespace:"trident", SelfLink:"", UID:"2f638a5c-7e38-11e8-a286-005056383598", ResourceVersion:"149707", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666160841, loc:(*time.Location)(0x1ec6320)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"trident", "projectcalico.org/orchestrator":"k8s"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"pike", ContainerID:"d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1", Pod:"pod2c", Endpoint:"eth0", IPNetworks:[]string{"192.168.0.55/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.trident"}, InterfaceName:"cali804c3f63559", MAC:"c2:67:46:d4:58:6f", Ports:[]v3.EndpointPort(nil)}} Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.190198 42179 config.go:297] Setting pods for source api Jul 02 16:40:42 pike kubelet[42179]: 2018-07-02 16:40:42.192 [INFO][53916] k8s.go 302: Wrote updated endpoint to datastore ContainerID="d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" Namespace="trident" Pod="pod2c" WorkloadEndpoint="pike-k8s-pod2c-eth0" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.208211 42179 kubelet.go:1863] SyncLoop (UPDATE, "api"): "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.211046 42179 kuberuntime_manager.go:654] Created PodSandbox "d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1" for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.212892 42179 prober.go:150] Exec-Probe Pod: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:trident-csi-0,GenerateName:trident-csi-,Namespace:trident,SelfLink:/api/v1/namespaces/trident/pods/trident-csi-0,UID:8eaa19b4-7e34-11e8-a286-005056383598,ResourceVersion:147708,Generation:0,CreationTimestamp:2018-07-02 16:14:43 -0400 EDT,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{app: controller.csi.trident.netapp.io,controller-revision-hash: trident-csi-66f8f9d5bb,statefulset.kubernetes.io/pod-name: trident-csi-0,},Annotations:map[string]string{cni.projectcalico.org/podIP: 192.168.0.49/32,kubernetes.io/config.seen: 2018-07-02T16:23:31.650583191-04:00,kubernetes.io/config.source: api,},OwnerReferences:[{apps/v1 StatefulSet trident-csi 8e98152e-7e34-11e8-a286-005056383598 0xc420438d08 0xc420438d09}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{etcd-vol {nil nil nil nil nil nil nil nil nil PersistentVolumeClaimVolumeSource{ClaimName:trident-csi,ReadOnly:false,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {socket-dir {nil &EmptyDirVolumeSource{Medium:,SizeLimit:,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {etc-dir {&HostPathVolumeSource{Path:/etc,Type:*Directory,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {trident-csi-token-4gzh4 {nil nil nil nil nil &SecretVolumeSource{SecretName:trident-csi-token-4gzh4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{trident-main registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae [/usr/local/bin/trident_orchestrator] [-etcd_v3 http://127.0.0.1:8001 --csi_node_name=$(KUBE_NODE_NAME) --csi_endpoint=$(CSI_ENDPOINT) -debug] [] [] [{KUBE_NODE_NAME EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldP Jul 02 16:40:42 pike kubelet[42179]: ath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,}} {CSI_ENDPOINT unix://plugin/csi.sock nil}] {map[] map[]} [{socket-dir false /plugin } {etc-dir false /etc } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[tridentctl -s 127.0.0.1:8000 get backend],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:120,TimeoutSeconds:90,PeriodSeconds:120,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false} {etcd quay.io/coreos/etcd:v3.2.19 [/usr/local/bin/etcd] [-name etcd1 -advertise-client-urls http://127.0.0.1:8001 -listen-client-urls http://127.0.0.1:8001 -initial-advertise-peer-urls http://127.0.0.1:8002 -listen-peer-urls http://127.0.0.1:8002 -data-dir /var/etcd/data -initial-cluster etcd1=http://127.0.0.1:8002] [] [] [] {map[] map[]} [{etcd-vol false /var/etcd/data } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:10,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false} {csi-attacher quay.io/k8scsi/csi-attacher:v0.2.0 [] [--v=9 --csi-address=$(ADDRESS)] [] [] [{ADDRESS /var/lib/csi/sockets/pluginproxy/csi.sock nil}] {map[] map[]} [{socket-dir false /var/lib/csi/sockets/pluginproxy/ } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false} {csi-provisioner quay.io/k8scsi/csi-provisioner:v0.2.1 [] [--v=9 --provisioner=io.netapp.trident.csi --csi-address=$(ADDRESS)] [] [] [{ADDRESS /var/lib/csi/sockets/pluginproxy/csi.sock nil}] {map[] map[]} [{socket-dir false /var/lib/csi/sockets/pluginproxy/ } {trident-csi-token-4gzh4 true /var/run/s Jul 02 16:40:42 pike kubelet[42179]: ecrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:trident-csi,DeprecatedServiceAccount:trident-csi,NodeName:pike,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,},ImagePullSecrets:[],Hostname:trident-csi-0,Subdomain:trident-csi,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc420438d18} {node.kubernetes.io/unreachable Exists NoExecute 0xc420438d20}],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:43 -0400 EDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:48 -0400 EDT } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-02 16:14:43 -0400 EDT }],Message:,Reason:,HostIP:192.168.228.17,PodIP:192.168.0.49,StartTime:2018-07-02 16:14:43 -0400 EDT,ContainerStatuses:[{csi-attacher {nil ContainerStateRunning{StartedAt:2018-07-02 16:14:47 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/k8scsi/csi-attacher:v0.2.0 docker-pullable://quay.io/k8scsi/csi-attacher@sha256:5cbb7934bd86d400c221379cff8b24ed4c06e121ea59608cfd7e67690100ba54 docker://f547da90c65d01207a828ee3b8d044eb4b4cc05fa065f39648eac3e539f3b9a6} {csi-provisioner {nil &ContainerStateRunning{StartedAt:2018-07-02 16:14:48 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/k8scsi/csi-provisioner:v0.2.1 docker-pullable://quay.io/k8scsi/csi-provisioner@sha256:fd4ed32315e846b6654f97c95b373da001fd9638cd5935c20a5bf9f5889e8602 docker://08d31d6c565ec486d8e9c46017c10edd0f390da6a139692a9daa25e620777bc0} {etcd {nil &ContainerStateRunning{StartedAt:2018-0 Jul 02 16:40:42 pike kubelet[42179]: 7-02 16:14:47 -0400 EDT,} nil} {nil nil nil} true 0 quay.io/coreos/etcd:v3.2.19 docker-pullable://quay.io/coreos/etcd@sha256:e21b9734136122b1e8b9b37742e72d2bd8b0d84bca875892494e0de5520313b0 docker://8340342e4a7288ef0672e62be7c0bf34e1800f7ce961607fa163146aa828b861} {trident-main {nil &ContainerStateRunning{StartedAt:2018-07-02 16:14:46 -0400 EDT,} nil} {nil nil nil} true 0 registry.barnacle.netapp.com:5000/trident:18.07.0-test.47be939f0c5dd5eb686af4937ce84c29ac2195ae docker-pullable://registry.barnacle.netapp.com:5000/trident@sha256:ce8cd402f9ef85e18ce428f09137a7347c5009c255daf730f9c40c006ea0ee23 docker://2c51e41d415063f9be931cf9ebbb62bf4d4c427d9002f312fbf54f9c34a4d9ee}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}, Container: {etcd quay.io/coreos/etcd:v3.2.19 [/usr/local/bin/etcd] [-name etcd1 -advertise-client-urls http://127.0.0.1:8001 -listen-client-urls http://127.0.0.1:8001 -initial-advertise-peer-urls http://127.0.0.1:8002 -listen-peer-urls http://127.0.0.1:8002 -data-dir /var/etcd/data -initial-cluster etcd1=http://127.0.0.1:8002] [] [] [] {map[] map[]} [{etcd-vol false /var/etcd/data } {trident-csi-token-4gzh4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] &Probe{Handler:Handler{Exec:&ExecAction{Command:[etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health],},HTTPGet:nil,TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:10,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:2,} nil nil /dev/termination-log File IfNotPresent nil false false false}, Command: [etcdctl -endpoint=http://127.0.0.1:8001/ cluster-health] Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.216926 42179 kuberuntime_manager.go:673] Determined the ip "192.168.0.55" for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" after sandbox changed Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.217008 42179 kuberuntime_manager.go:724] Creating container &Container{Name:nginx,Image:nginx:latest,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TCP }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{data false /var/www } {default-token-9dv6w true /var/run/secrets/kubernetes.io/serviceaccount }],LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,VolumeDevices:[],} in pod pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598) Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.220106 42179 kuberuntime_image.go:47] Pulling image "nginx:latest" without credentials Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.220470 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx:latest" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.286593 42179 exec.go:38] Exec probe response: "member 9ca0c7938a0c3dc5 is healthy: got healthy result from http://127.0.0.1:8001\ncluster is healthy\n" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.286616 42179 prober.go:118] Liveness probe for "trident-csi-0_trident(8eaa19b4-7e34-11e8-a286-005056383598):etcd" succeeded Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.639687 42179 generic.go:147] GenericPLEG: 2f638a5c-7e38-11e8-a286-005056383598/d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1: non-existent -> running Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.641132 42179 kuberuntime_manager.go:852] getSandboxIDByPodUID got sandbox IDs ["d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1"] for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.645170 42179 generic.go:380] PLEG: Write status for pod2c/trident: &container.PodStatus{ID:"2f638a5c-7e38-11e8-a286-005056383598", Name:"pod2c", Namespace:"trident", IP:"192.168.0.55", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*v1alpha2.PodSandboxStatus{(*v1alpha2.PodSandboxStatus)(0xc420e1f450)}} (err: ) Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.645241 42179 kubelet.go:1901] SyncLoop (PLEG): "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)", event: &pleg.PodLifecycleEvent{ID:"2f638a5c-7e38-11e8-a286-005056383598", Type:"ContainerStarted", Data:"d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1"} Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.817273 42179 kube_docker_client.go:348] Stop pulling image "nginx:latest": "Status: Image is up to date for nginx:latest" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.819064 42179 kuberuntime_container.go:100] Generating ref for container nginx: &v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:"spec.containers{nginx}"} Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.819191 42179 kubelet_pods.go:172] container: trident/pod2c/nginx podIP: "192.168.0.55" creating hosts mount: true Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.819226 42179 csi_mounter.go:77] kubernetes.io/csi: mounter.GetPath generated [/var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~csi/pvc-281709ce7e3711e8/mount] Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.819389 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "nginx:latest" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.821947 42179 docker_service.go:440] Setting cgroup parent to: "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598" Jul 02 16:40:42 pike kubelet[42179]: I0702 16:40:42.943310 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Created' Created container Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.114419 42179 factory.go:113] Using factory "docker" for container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.115972 42179 server.go:428] Event(v1.ObjectReference{Kind:"Pod", Namespace:"trident", Name:"pod2c", UID:"2f638a5c-7e38-11e8-a286-005056383598", APIVersion:"v1", ResourceVersion:"149705", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Started' Started container Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.117534 42179 manager.go:997] Added container: "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318" (aliases: [k8s_nginx_pod2c_trident_2f638a5c-7e38-11e8-a286-005056383598_0 f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318], namespace: "docker") Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.117785 42179 handler.go:325] Added event &{/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318 2018-07-02 20:40:42.826042577 +0000 UTC containerCreation {}} Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.117822 42179 container.go:448] Start housekeeping for container "/kubepods/besteffort/pod2f638a5c-7e38-11e8-a286-005056383598/f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.269657 42179 interface.go:360] Looking for default routes with IPv4 addresses Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.269705 42179 interface.go:365] Default route transits interface "ens33" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.270838 42179 interface.go:174] Interface ens33 is up Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.271020 42179 interface.go:222] Interface "ens33" has 2 addresses :[192.168.228.17/24 fe80::779f:8935:8254:8bd8/64]. Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.271047 42179 interface.go:189] Checking addr 192.168.228.17/24. Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.271079 42179 interface.go:196] IP found 192.168.228.17 Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.271089 42179 interface.go:228] Found valid IPv4 address 192.168.228.17 for interface "ens33". Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.271098 42179 interface.go:371] Found active IP 192.168.228.17 Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.389107 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /readiness Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.389466 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.390664 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/readiness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:43 GMT]] 0xc4217c4380 0 [] true false map[] 0xc422906900 } Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.390931 42179 prober.go:118] Readiness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.656659 42179 generic.go:147] GenericPLEG: 2f638a5c-7e38-11e8-a286-005056383598/f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318: non-existent -> running Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.658424 42179 kuberuntime_manager.go:852] getSandboxIDByPodUID got sandbox IDs ["d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1"] for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.665848 42179 generic.go:380] PLEG: Write status for pod2c/trident: &container.PodStatus{ID:"2f638a5c-7e38-11e8-a286-005056383598", Name:"pod2c", Namespace:"trident", IP:"192.168.0.55", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc420baca80)}, SandboxStatuses:[]*v1alpha2.PodSandboxStatus{(*v1alpha2.PodSandboxStatus)(0xc422f1a4b0)}} (err: ) Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.666161 42179 kubelet.go:1901] SyncLoop (PLEG): "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)", event: &pleg.PodLifecycleEvent{ID:"2f638a5c-7e38-11e8-a286-005056383598", Type:"ContainerStarted", Data:"f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318"} Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.666219 42179 kubelet_pods.go:1366] Generating status for "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.666691 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.674033 42179 config.go:297] Setting pods for source api Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.677338 42179 status_manager.go:481] Status for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:43 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.55 StartTime:2018-07-02 16:40:41 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:40:43 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:nginx:latest ImageID:docker-pullable://nginx@sha256:62a095e5da5f977b9f830adaf64d604c614024bf239d21068e4ca826d0d629a4 ContainerID:docker://f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318}] QOSClass:BestEffort}) Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.678869 42179 kubelet.go:1869] SyncLoop (RECONCILE, "api"): "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.699987 42179 kubelet.go:1943] SyncLoop (housekeeping) Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.721556 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "2f638a5c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.721615 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "2f638a5c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.724742 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.725061 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 2f638a5c-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.727368 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.727619 42179 atomic_writer.go:156] pod trident/pod2c volume default-token-9dv6w: no update required for target directory /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.727641 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.967106 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:43 pike kubelet[42179]: I0702 16:40:43.967210 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.144887 42179 prober.go:165] HTTP-Probe Host: http://192.168.228.17, Port: 9099, Path: /liveness Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.144981 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.146446 42179 http.go:96] Probe succeeded for http://192.168.228.17:9099/liveness, Response: {204 No Content 204 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:44 GMT]] 0xc421153ca0 0 [] true false map[] 0xc422b62100 } Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.146537 42179 prober.go:118] Liveness probe for "calico-node-fnnvb_kube-system(fe1240e4-773a-11e8-a95f-005056383598):calico-node" succeeded Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.614356 42179 prober.go:165] HTTP-Probe Host: http://192.168.0.43, Port: 10054, Path: /metrics Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.614462 42179 prober.go:168] HTTP-Probe Headers: map[] Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.624780 42179 http.go:96] Probe succeeded for http://192.168.0.43:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Mon, 02 Jul 2018 20:40:44 GMT] Content-Type:[text/plain; version=0.0.4]] 0xc4206a19c0 -1 [] true true map[] 0xc4224deb00 } Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.624849 42179 prober.go:118] Liveness probe for "kube-dns-86f4d74b45-6hhgk_kube-system(bafaa6c3-773a-11e8-a95f-005056383598):sidecar" succeeded Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.681389 42179 kubelet_pods.go:1366] Generating status for "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.681570 42179 status_manager.go:340] Ignoring same status for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:43 -0400 EDT Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-02 16:40:41 -0400 EDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.228.17 PodIP:192.168.0.55 StartTime:2018-07-02 16:40:41 -0400 EDT InitContainerStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2018-07-02 16:40:43 -0400 EDT,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:nginx:latest ImageID:docker-pullable://nginx@sha256:62a095e5da5f977b9f830adaf64d604c614024bf239d21068e4ca826d0d629a4 ContainerID:docker://f7b2795dcbf527679c38259db9fe38ec36dfad12cc3042e9b2ac34c0afd46318}] QOSClass:BestEffort} Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.681705 42179 volume_manager.go:347] Waiting for volumes to attach and mount for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.686737 42179 kubelet.go:2122] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message: Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.733964 42179 desired_state_of_world_populator.go:302] Added volume "data" (volSpec="pvc-281709ce7e3711e8") for pod "2f638a5c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.734046 42179 desired_state_of_world_populator.go:302] Added volume "default-token-9dv6w" (volSpec="default-token-9dv6w") for pod "2f638a5c-7e38-11e8-a286-005056383598" to desired state. Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.831662 42179 reconciler.go:237] Starting operationExecutor.MountVolume for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Volume is already mounted to pod, but remount was requested. Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.831847 42179 secret.go:186] Setting up volume default-token-9dv6w for pod 2f638a5c-7e38-11e8-a286-005056383598 at /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.833515 42179 secret.go:216] Received secret trident/default-token-9dv6w containing (3) pieces of data, 1874 total bytes Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.833844 42179 atomic_writer.go:156] pod trident/pod2c volume default-token-9dv6w: no update required for target directory /var/lib/kubelet/pods/2f638a5c-7e38-11e8-a286-005056383598/volumes/kubernetes.io~secret/default-token-9dv6w Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.834042 42179 operation_generator.go:557] MountVolume.SetUp succeeded for volume "default-token-9dv6w" (UniqueName: "kubernetes.io/secret/2f638a5c-7e38-11e8-a286-005056383598-default-token-9dv6w") pod "pod2c" (UID: "2f638a5c-7e38-11e8-a286-005056383598") Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.981973 42179 volume_manager.go:380] All volumes are attached and mounted for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:44 pike kubelet[42179]: I0702 16:40:44.982101 42179 kuberuntime_manager.go:570] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d51f41837e66f87c1dcf0906e1c7544b4dce888c0537d043cd2626848a8d31d1 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "pod2c_trident(2f638a5c-7e38-11e8-a286-005056383598)" Jul 02 16:40:45 pike kubelet[42179]: I0702 16:40:45.699885 42179 kubelet.go:1943] SyncLoop (housekeeping)