Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- oc describe pod rook-ceph-osd-0-75765667f8-pl4c9 -n openshift-storage
- Name: rook-ceph-osd-0-75765667f8-pl4c9
- Namespace: openshift-storage
- Priority: 2000001000
- Priority Class Name: system-node-critical
- Node: rhocs-bm8.lab.eng.blr.redhat.com/10.70.61.8
- Start Time: Mon, 07 Jun 2021 15:44:04 +0530
- Labels: app=rook-ceph-osd
- ceph-osd-id=0
- ceph-version=16.2.4-0
- ceph.rook.io/DeviceSet=example-deviceset-1
- ceph.rook.io/pvc=example-deviceset-1-data-0plqzb
- ceph_daemon_id=0
- ceph_daemon_type=osd
- failure-domain=rhocs-bm8.lab.eng.blr.redhat.com
- osd=0
- pod-template-hash=75765667f8
- portable=false
- rook-version=v1.6.3
- rook_cluster=openshift-storage
- topology-location-host=rhocs-bm8-lab-eng-blr-redhat-com
- topology-location-rack=rack0
- topology-location-root=default
- Annotations: k8s.v1.cni.cncf.io/network-status:
- [{
- "name": "",
- "interface": "eth0",
- "ips": [
- "10.130.0.251"
- ],
- "default": true,
- "dns": {}
- },{
- "name": "openshift-storage/ipv6",
- "interface": "net1",
- "ips": [
- "2620:52:0:463e:22e8:b8fb:89ae:3813"
- ],
- "mac": "ce:0c:f3:9f:bd:2c",
- "dns": {}
- },{
- "name": "openshift-storage/ipv6",
- "interface": "net2",
- "ips": [
- "2620:52:0:463e:22e8:b8fb:89ae:3814"
- ],
- "mac": "de:2e:9a:50:bc:14",
- "dns": {}
- }]
- k8s.v1.cni.cncf.io/networks: ipv6, ipv6
- k8s.v1.cni.cncf.io/networks-status:
- [{
- "name": "",
- "interface": "eth0",
- "ips": [
- "10.130.0.251"
- ],
- "default": true,
- "dns": {}
- },{
- "name": "openshift-storage/ipv6",
- "interface": "net1",
- "ips": [
- "2620:52:0:463e:22e8:b8fb:89ae:3813"
- ],
- "mac": "ce:0c:f3:9f:bd:2c",
- "dns": {}
- },{
- "name": "openshift-storage/ipv6",
- "interface": "net2",
- "ips": [
- "2620:52:0:463e:22e8:b8fb:89ae:3814"
- ],
- "mac": "de:2e:9a:50:bc:14",
- "dns": {}
- }]
- openshift.io/scc: rook-ceph
- Status: Running
- IP: 10.130.0.251
- IPs:
- IP: 10.130.0.251
- Controlled By: ReplicaSet/rook-ceph-osd-0-75765667f8
- Init Containers:
- blkdevmapper:
- Container ID: cri-o://e3e2730e4f5c434404abd8100270e111afcc8ac05d49819a7ee1587eaf81471c
- Image: ceph/daemon-base:latest-pacific
- Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
- Port: <none>
- Host Port: <none>
- Command:
- /bin/bash
- -c
- set -xe
- PVC_SOURCE=/example-deviceset-1-data-0plqzb
- PVC_DEST=/var/lib/ceph/osd/ceph-0/block
- CP_ARGS=(--archive --dereference --verbose)
- if [ -b "$PVC_DEST" ]; then
- PVC_SOURCE_MAJ_MIN=$(stat --format '%t%T' $PVC_SOURCE)
- PVC_DEST_MAJ_MIN=$(stat --format '%t%T' $PVC_DEST)
- if [[ "$PVC_SOURCE_MAJ_MIN" == "$PVC_DEST_MAJ_MIN" ]]; then
- CP_ARGS+=(--no-clobber)
- else
- echo "PVC's source major/minor numbers changed"
- CP_ARGS+=(--remove-destination)
- fi
- fi
- cp "${CP_ARGS[@]}" "$PVC_SOURCE" "$PVC_DEST"
- State: Terminated
- Reason: Completed
- Exit Code: 0
- Started: Mon, 07 Jun 2021 15:44:08 +0530
- Finished: Mon, 07 Jun 2021 15:44:08 +0530
- Ready: True
- Restart Count: 0
- Limits:
- cpu: 2
- memory: 5Gi
- Requests:
- cpu: 2
- memory: 5Gi
- Environment: <none>
- Mounts:
- /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
- /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
- Devices:
- /example-deviceset-1-data-0plqzb from example-deviceset-1-data-0plqzb
- activate:
- Container ID: cri-o://8ce2751fd873982901c02d9aab70dd7f278107f27082723b8a53768a38d5bf98
- Image: ceph/daemon-base:latest-pacific
- Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
- Port: <none>
- Host Port: <none>
- Command:
- ceph-bluestore-tool
- Args:
- prime-osd-dir
- --dev
- /var/lib/ceph/osd/ceph-0/block
- --path
- /var/lib/ceph/osd/ceph-0
- --no-mon-config
- State: Terminated
- Reason: Completed
- Exit Code: 0
- Started: Mon, 07 Jun 2021 15:44:08 +0530
- Finished: Mon, 07 Jun 2021 15:44:08 +0530
- Ready: True
- Restart Count: 0
- Limits:
- cpu: 2
- memory: 5Gi
- Requests:
- cpu: 2
- memory: 5Gi
- Environment: <none>
- Mounts:
- /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
- /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
- Devices:
- /var/lib/ceph/osd/ceph-0/block from example-deviceset-1-data-0plqzb
- expand-bluefs:
- Container ID: cri-o://b89d063beba95ce38ed44bdc2a00b9734871457e51bca43506c1741408b7821c
- Image: ceph/daemon-base:latest-pacific
- Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
- Port: <none>
- Host Port: <none>
- Command:
- ceph-bluestore-tool
- Args:
- bluefs-bdev-expand
- --path
- /var/lib/ceph/osd/ceph-0
- State: Terminated
- Reason: Completed
- Exit Code: 0
- Started: Mon, 07 Jun 2021 15:44:09 +0530
- Finished: Mon, 07 Jun 2021 15:44:10 +0530
- Ready: True
- Restart Count: 0
- Limits:
- cpu: 2
- memory: 5Gi
- Requests:
- cpu: 2
- memory: 5Gi
- Environment: <none>
- Mounts:
- /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
- /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
- chown-container-data-dir:
- Container ID: cri-o://35943ef7c4db52c5100ac66159ae3b0131ebef17ae018e313d97dbec36b380cc
- Image: ceph/daemon-base:latest-pacific
- Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
- Port: <none>
- Host Port: <none>
- Command:
- chown
- Args:
- --verbose
- --recursive
- ceph:ceph
- /var/log/ceph
- /var/lib/ceph/crash
- /var/lib/ceph/osd/ceph-0
- State: Terminated
- Reason: Completed
- Exit Code: 0
- Started: Mon, 07 Jun 2021 15:44:11 +0530
- Finished: Mon, 07 Jun 2021 15:44:11 +0530
- Ready: True
- Restart Count: 0
- Limits:
- cpu: 2
- memory: 5Gi
- Requests:
- cpu: 2
- memory: 5Gi
- Environment: <none>
- Mounts:
- /etc/ceph from rook-config-override (ro)
- /run/udev from run-udev (rw)
- /var/lib/ceph/crash from rook-ceph-crash (rw)
- /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
- /var/lib/rook from rook-data (rw)
- /var/log/ceph from rook-ceph-log (rw)
- /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
- Containers:
- osd:
- Container ID: cri-o://8e67badb674c83f7f8506f61fd9ce191eafece9b981bc24bec6325b07997e43a
- Image: ceph/daemon-base:latest-pacific
- Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
- Port: <none>
- Host Port: <none>
- Command:
- ceph-osd
- Args:
- --foreground
- --id
- 0
- --fsid
- 36c7dde1-761c-44a0-bfed-af3b7b51181f
- --setuser
- ceph
- --setgroup
- ceph
- --crush-location=root=default host=rhocs-bm8-lab-eng-blr-redhat-com rack=rack0
- --log-to-stderr=true
- --err-to-stderr=true
- --mon-cluster-log-to-stderr=true
- --log-stderr-prefix=debug
- --default-log-to-file=false
- --default-mon-cluster-log-to-file=false
- --ms-learn-addr-from-peer=false
- --ms-bind-ipv4=true
- --ms-bind-ipv6=true
- State: Waiting
- Reason: CrashLoopBackOff
- Last State: Terminated
- Reason: Error
- Exit Code: 1
- Started: Mon, 07 Jun 2021 17:32:56 +0530
- Finished: Mon, 07 Jun 2021 17:32:57 +0530
- Ready: False
- Restart Count: 26
- Limits:
- cpu: 2
- memory: 5Gi
- Requests:
- cpu: 2
- memory: 5Gi
- Liveness: exec [env -i sh -c ceph --admin-daemon /run/ceph/ceph-osd.0.asok status] delay=45s timeout=1s period=10s #success=1 #failure=3
- Environment:
- ROOK_NODE_NAME: rhocs-bm8.lab.eng.blr.redhat.com
- ROOK_CLUSTER_ID: 0090b395-f582-4077-b36d-838db2b8b6f1
- ROOK_CLUSTER_NAME: ocs-storagecluster-cephcluster
- ROOK_PRIVATE_IP: (v1:status.podIP)
- ROOK_PUBLIC_IP: (v1:status.podIP)
- POD_NAMESPACE: openshift-storage
- ROOK_MON_ENDPOINTS: <set to the key 'data' of config map 'rook-ceph-mon-endpoints'> Optional: false
- ROOK_MON_SECRET: <set to the key 'mon-secret' in secret 'rook-ceph-mon'> Optional: false
- ROOK_CEPH_USERNAME: <set to the key 'ceph-username' in secret 'rook-ceph-mon'> Optional: false
- ROOK_CEPH_SECRET: <set to the key 'ceph-secret' in secret 'rook-ceph-mon'> Optional: false
- ROOK_CONFIG_DIR: /var/lib/rook
- ROOK_CEPH_CONFIG_OVERRIDE: /etc/rook/config/override.conf
- ROOK_FSID: <set to the key 'fsid' in secret 'rook-ceph-mon'> Optional: false
- NODE_NAME: (v1:spec.nodeName)
- ROOK_CRUSHMAP_ROOT: default
- ROOK_CRUSHMAP_HOSTNAME:
- CEPH_VOLUME_DEBUG: 1
- CEPH_VOLUME_SKIP_RESTORECON: 1
- DM_DISABLE_UDEV: 1
- TINI_SUBREAPER:
- CONTAINER_IMAGE: ceph/daemon-base:latest-pacific
- POD_NAME: rook-ceph-osd-0-75765667f8-pl4c9 (v1:metadata.name)
- POD_MEMORY_LIMIT: 5368709120 (limits.memory)
- POD_MEMORY_REQUEST: 5368709120 (requests.memory)
- POD_CPU_LIMIT: 2 (limits.cpu)
- POD_CPU_REQUEST: 2 (requests.cpu)
- ROOK_OSD_UUID: 1fa42919-fc32-4ab5-a06a-e2c93b33c64a
- ROOK_OSD_ID: 0
- ROOK_CEPH_MON_HOST: <set to the key 'mon_host' in secret 'rook-ceph-config'> Optional: false
- CEPH_ARGS: -m $(ROOK_CEPH_MON_HOST)
- ROOK_BLOCK_PATH: /mnt/example-deviceset-1-data-0plqzb
- ROOK_CV_MODE: raw
- ROOK_OSD_PVC_SIZE: 100Gi
- ROOK_PVC_BACKED_OSD: true
- Mounts:
- /etc/ceph from rook-config-override (ro)
- /run/udev from run-udev (rw)
- /var/lib/ceph/crash from rook-ceph-crash (rw)
- /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
- /var/lib/rook from rook-data (rw)
- /var/log/ceph from rook-ceph-log (rw)
- /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
- log-collector:
- Container ID: cri-o://fbc480fd8763a76417f618ca6cad0f822dcba2efa801fdcc5696a136ff18672b
- Image: ceph/daemon-base:latest-pacific
- Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
- Port: <none>
- Host Port: <none>
- Command:
- /bin/bash
- -c
- set -xe
- CEPH_CLIENT_ID=ceph-osd.0
- PERIODICITY=24h
- LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph
- if [ -z "$PERIODICITY" ]; then
- PERIODICITY=24h
- fi
- # edit the logrotate file to only rotate a specific daemon log
- # otherwise we will logrotate log files without reloading certain daemons
- # this might happen when multiple daemons run on the same machine
- sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE"
- while true; do
- sleep "$PERIODICITY"
- echo "starting log rotation"
- logrotate --verbose --force "$LOG_ROTATE_CEPH_FILE"
- echo "I am going to sleep now, see you in $PERIODICITY"
- done
- State: Running
- Started: Mon, 07 Jun 2021 15:44:13 +0530
- Ready: True
- Restart Count: 0
- Environment: <none>
- Mounts:
- /etc/ceph from rook-config-override (ro)
- /var/lib/ceph/crash from rook-ceph-crash (rw)
- /var/log/ceph from rook-ceph-log (rw)
- /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
- Conditions:
- Type Status
- Initialized True
- Ready False
- ContainersReady False
- PodScheduled True
- Volumes:
- rook-data:
- Type: EmptyDir (a temporary directory that shares a pod's lifetime)
- Medium:
- SizeLimit: <unset>
- rook-config-override:
- Type: Projected (a volume that contains injected data from multiple sources)
- ConfigMapName: rook-config-override
- ConfigMapOptional: <nil>
- rook-ceph-log:
- Type: HostPath (bare host directory volume)
- Path: /var/lib/rook/openshift-storage/log
- HostPathType:
- rook-ceph-crash:
- Type: HostPath (bare host directory volume)
- Path: /var/lib/rook/openshift-storage/crash
- HostPathType:
- example-deviceset-1-data-0plqzb:
- Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace)
- ClaimName: example-deviceset-1-data-0plqzb
- ReadOnly: false
- example-deviceset-1-data-0plqzb-bridge:
- Type: HostPath (bare host directory volume)
- Path: /var/lib/rook/openshift-storage/example-deviceset-1-data-0plqzb
- HostPathType: DirectoryOrCreate
- run-udev:
- Type: HostPath (bare host directory volume)
- Path: /run/udev
- HostPathType:
- rook-ceph-osd-token-qljzb:
- Type: Secret (a volume populated by a Secret)
- SecretName: rook-ceph-osd-token-qljzb
- Optional: false
- QoS Class: Burstable
- Node-Selectors: kubernetes.io/hostname=rhocs-bm8.lab.eng.blr.redhat.com
- Tolerations: node.kubernetes.io/memory-pressure:NoSchedule op=Exists
- node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
- node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
- node.ocs.openshift.io/storage=true:NoSchedule
- Events:
- Type Reason Age From Message
- ---- ------ ---- ---- -------
- Normal Scheduled 108m default-scheduler Successfully assigned openshift-storage/rook-ceph-osd-0-75765667f8-pl4c9 to rhocs-bm8.lab.eng.blr.redhat.com
- Normal SuccessfulMountVolume 108m kubelet MapVolume.MapPodDevice succeeded for volume "local-pv-6b183b4d" globalMapPath "/var/lib/kubelet/plugins/kubernetes.io~local-volume/volumeDevices/local-pv-6b183b4d"
- Normal SuccessfulMountVolume 108m kubelet MapVolume.MapPodDevice succeeded for volume "local-pv-6b183b4d" volumeMapPath "/var/lib/kubelet/pods/88e797db-e8bc-42fa-a5cd-2e1f9ff5ac55/volumeDevices/kubernetes.io~local-volume"
- Normal AddedInterface 108m multus Add eth0 [10.130.0.251/23]
- Normal AddedInterface 108m multus Add net1 [2620:52:0:463e:22e8:b8fb:89ae:3813/117] from openshift-storage/ipv6
- Normal AddedInterface 108m multus Add net2 [2620:52:0:463e:22e8:b8fb:89ae:3814/117] from openshift-storage/ipv6
- Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
- Normal Created 108m kubelet Created container blkdevmapper
- Normal Started 108m kubelet Started container blkdevmapper
- Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
- Normal Created 108m kubelet Created container activate
- Normal Started 108m kubelet Started container activate
- Normal Started 108m kubelet Started container expand-bluefs
- Normal Created 108m kubelet Created container expand-bluefs
- Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
- Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
- Normal Created 108m kubelet Created container chown-container-data-dir
- Normal Started 108m kubelet Started container chown-container-data-dir
- Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
- Normal Pulled 108m (x2 over 108m) kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
- Normal Created 108m (x2 over 108m) kubelet Created container osd
- Normal Started 108m (x2 over 108m) kubelet Started container osd
- Normal Created 108m kubelet Created container log-collector
- Normal Started 108m kubelet Started container log-collector
- Warning BackOff 3m43s (x503 over 108m) kubelet Back-off restarting failed container
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement