Advertisement
kmajumder

osd_describe

Jun 7th, 2021
82
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 19.83 KB | None | 0 0
  1. oc describe pod rook-ceph-osd-0-75765667f8-pl4c9 -n openshift-storage
  2. Name: rook-ceph-osd-0-75765667f8-pl4c9
  3. Namespace: openshift-storage
  4. Priority: 2000001000
  5. Priority Class Name: system-node-critical
  6. Node: rhocs-bm8.lab.eng.blr.redhat.com/10.70.61.8
  7. Start Time: Mon, 07 Jun 2021 15:44:04 +0530
  8. Labels: app=rook-ceph-osd
  9. ceph-osd-id=0
  10. ceph-version=16.2.4-0
  11. ceph.rook.io/DeviceSet=example-deviceset-1
  12. ceph.rook.io/pvc=example-deviceset-1-data-0plqzb
  13. ceph_daemon_id=0
  14. ceph_daemon_type=osd
  15. failure-domain=rhocs-bm8.lab.eng.blr.redhat.com
  16. osd=0
  17. pod-template-hash=75765667f8
  18. portable=false
  19. rook-version=v1.6.3
  20. rook_cluster=openshift-storage
  21. topology-location-host=rhocs-bm8-lab-eng-blr-redhat-com
  22. topology-location-rack=rack0
  23. topology-location-root=default
  24. Annotations: k8s.v1.cni.cncf.io/network-status:
  25. [{
  26. "name": "",
  27. "interface": "eth0",
  28. "ips": [
  29. "10.130.0.251"
  30. ],
  31. "default": true,
  32. "dns": {}
  33. },{
  34. "name": "openshift-storage/ipv6",
  35. "interface": "net1",
  36. "ips": [
  37. "2620:52:0:463e:22e8:b8fb:89ae:3813"
  38. ],
  39. "mac": "ce:0c:f3:9f:bd:2c",
  40. "dns": {}
  41. },{
  42. "name": "openshift-storage/ipv6",
  43. "interface": "net2",
  44. "ips": [
  45. "2620:52:0:463e:22e8:b8fb:89ae:3814"
  46. ],
  47. "mac": "de:2e:9a:50:bc:14",
  48. "dns": {}
  49. }]
  50. k8s.v1.cni.cncf.io/networks: ipv6, ipv6
  51. k8s.v1.cni.cncf.io/networks-status:
  52. [{
  53. "name": "",
  54. "interface": "eth0",
  55. "ips": [
  56. "10.130.0.251"
  57. ],
  58. "default": true,
  59. "dns": {}
  60. },{
  61. "name": "openshift-storage/ipv6",
  62. "interface": "net1",
  63. "ips": [
  64. "2620:52:0:463e:22e8:b8fb:89ae:3813"
  65. ],
  66. "mac": "ce:0c:f3:9f:bd:2c",
  67. "dns": {}
  68. },{
  69. "name": "openshift-storage/ipv6",
  70. "interface": "net2",
  71. "ips": [
  72. "2620:52:0:463e:22e8:b8fb:89ae:3814"
  73. ],
  74. "mac": "de:2e:9a:50:bc:14",
  75. "dns": {}
  76. }]
  77. openshift.io/scc: rook-ceph
  78. Status: Running
  79. IP: 10.130.0.251
  80. IPs:
  81. IP: 10.130.0.251
  82. Controlled By: ReplicaSet/rook-ceph-osd-0-75765667f8
  83. Init Containers:
  84. blkdevmapper:
  85. Container ID: cri-o://e3e2730e4f5c434404abd8100270e111afcc8ac05d49819a7ee1587eaf81471c
  86. Image: ceph/daemon-base:latest-pacific
  87. Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
  88. Port: <none>
  89. Host Port: <none>
  90. Command:
  91. /bin/bash
  92. -c
  93.  
  94. set -xe
  95.  
  96. PVC_SOURCE=/example-deviceset-1-data-0plqzb
  97. PVC_DEST=/var/lib/ceph/osd/ceph-0/block
  98. CP_ARGS=(--archive --dereference --verbose)
  99.  
  100. if [ -b "$PVC_DEST" ]; then
  101. PVC_SOURCE_MAJ_MIN=$(stat --format '%t%T' $PVC_SOURCE)
  102. PVC_DEST_MAJ_MIN=$(stat --format '%t%T' $PVC_DEST)
  103. if [[ "$PVC_SOURCE_MAJ_MIN" == "$PVC_DEST_MAJ_MIN" ]]; then
  104. CP_ARGS+=(--no-clobber)
  105. else
  106. echo "PVC's source major/minor numbers changed"
  107. CP_ARGS+=(--remove-destination)
  108. fi
  109. fi
  110.  
  111. cp "${CP_ARGS[@]}" "$PVC_SOURCE" "$PVC_DEST"
  112.  
  113. State: Terminated
  114. Reason: Completed
  115. Exit Code: 0
  116. Started: Mon, 07 Jun 2021 15:44:08 +0530
  117. Finished: Mon, 07 Jun 2021 15:44:08 +0530
  118. Ready: True
  119. Restart Count: 0
  120. Limits:
  121. cpu: 2
  122. memory: 5Gi
  123. Requests:
  124. cpu: 2
  125. memory: 5Gi
  126. Environment: <none>
  127. Mounts:
  128. /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
  129. /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
  130. Devices:
  131. /example-deviceset-1-data-0plqzb from example-deviceset-1-data-0plqzb
  132. activate:
  133. Container ID: cri-o://8ce2751fd873982901c02d9aab70dd7f278107f27082723b8a53768a38d5bf98
  134. Image: ceph/daemon-base:latest-pacific
  135. Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
  136. Port: <none>
  137. Host Port: <none>
  138. Command:
  139. ceph-bluestore-tool
  140. Args:
  141. prime-osd-dir
  142. --dev
  143. /var/lib/ceph/osd/ceph-0/block
  144. --path
  145. /var/lib/ceph/osd/ceph-0
  146. --no-mon-config
  147. State: Terminated
  148. Reason: Completed
  149. Exit Code: 0
  150. Started: Mon, 07 Jun 2021 15:44:08 +0530
  151. Finished: Mon, 07 Jun 2021 15:44:08 +0530
  152. Ready: True
  153. Restart Count: 0
  154. Limits:
  155. cpu: 2
  156. memory: 5Gi
  157. Requests:
  158. cpu: 2
  159. memory: 5Gi
  160. Environment: <none>
  161. Mounts:
  162. /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
  163. /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
  164. Devices:
  165. /var/lib/ceph/osd/ceph-0/block from example-deviceset-1-data-0plqzb
  166. expand-bluefs:
  167. Container ID: cri-o://b89d063beba95ce38ed44bdc2a00b9734871457e51bca43506c1741408b7821c
  168. Image: ceph/daemon-base:latest-pacific
  169. Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
  170. Port: <none>
  171. Host Port: <none>
  172. Command:
  173. ceph-bluestore-tool
  174. Args:
  175. bluefs-bdev-expand
  176. --path
  177. /var/lib/ceph/osd/ceph-0
  178. State: Terminated
  179. Reason: Completed
  180. Exit Code: 0
  181. Started: Mon, 07 Jun 2021 15:44:09 +0530
  182. Finished: Mon, 07 Jun 2021 15:44:10 +0530
  183. Ready: True
  184. Restart Count: 0
  185. Limits:
  186. cpu: 2
  187. memory: 5Gi
  188. Requests:
  189. cpu: 2
  190. memory: 5Gi
  191. Environment: <none>
  192. Mounts:
  193. /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
  194. /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
  195. chown-container-data-dir:
  196. Container ID: cri-o://35943ef7c4db52c5100ac66159ae3b0131ebef17ae018e313d97dbec36b380cc
  197. Image: ceph/daemon-base:latest-pacific
  198. Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
  199. Port: <none>
  200. Host Port: <none>
  201. Command:
  202. chown
  203. Args:
  204. --verbose
  205. --recursive
  206. ceph:ceph
  207. /var/log/ceph
  208. /var/lib/ceph/crash
  209. /var/lib/ceph/osd/ceph-0
  210. State: Terminated
  211. Reason: Completed
  212. Exit Code: 0
  213. Started: Mon, 07 Jun 2021 15:44:11 +0530
  214. Finished: Mon, 07 Jun 2021 15:44:11 +0530
  215. Ready: True
  216. Restart Count: 0
  217. Limits:
  218. cpu: 2
  219. memory: 5Gi
  220. Requests:
  221. cpu: 2
  222. memory: 5Gi
  223. Environment: <none>
  224. Mounts:
  225. /etc/ceph from rook-config-override (ro)
  226. /run/udev from run-udev (rw)
  227. /var/lib/ceph/crash from rook-ceph-crash (rw)
  228. /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
  229. /var/lib/rook from rook-data (rw)
  230. /var/log/ceph from rook-ceph-log (rw)
  231. /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
  232. Containers:
  233. osd:
  234. Container ID: cri-o://8e67badb674c83f7f8506f61fd9ce191eafece9b981bc24bec6325b07997e43a
  235. Image: ceph/daemon-base:latest-pacific
  236. Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
  237. Port: <none>
  238. Host Port: <none>
  239. Command:
  240. ceph-osd
  241. Args:
  242. --foreground
  243. --id
  244. 0
  245. --fsid
  246. 36c7dde1-761c-44a0-bfed-af3b7b51181f
  247. --setuser
  248. ceph
  249. --setgroup
  250. ceph
  251. --crush-location=root=default host=rhocs-bm8-lab-eng-blr-redhat-com rack=rack0
  252. --log-to-stderr=true
  253. --err-to-stderr=true
  254. --mon-cluster-log-to-stderr=true
  255. --log-stderr-prefix=debug
  256. --default-log-to-file=false
  257. --default-mon-cluster-log-to-file=false
  258. --ms-learn-addr-from-peer=false
  259. --ms-bind-ipv4=true
  260. --ms-bind-ipv6=true
  261. State: Waiting
  262. Reason: CrashLoopBackOff
  263. Last State: Terminated
  264. Reason: Error
  265. Exit Code: 1
  266. Started: Mon, 07 Jun 2021 17:32:56 +0530
  267. Finished: Mon, 07 Jun 2021 17:32:57 +0530
  268. Ready: False
  269. Restart Count: 26
  270. Limits:
  271. cpu: 2
  272. memory: 5Gi
  273. Requests:
  274. cpu: 2
  275. memory: 5Gi
  276. Liveness: exec [env -i sh -c ceph --admin-daemon /run/ceph/ceph-osd.0.asok status] delay=45s timeout=1s period=10s #success=1 #failure=3
  277. Environment:
  278. ROOK_NODE_NAME: rhocs-bm8.lab.eng.blr.redhat.com
  279. ROOK_CLUSTER_ID: 0090b395-f582-4077-b36d-838db2b8b6f1
  280. ROOK_CLUSTER_NAME: ocs-storagecluster-cephcluster
  281. ROOK_PRIVATE_IP: (v1:status.podIP)
  282. ROOK_PUBLIC_IP: (v1:status.podIP)
  283. POD_NAMESPACE: openshift-storage
  284. ROOK_MON_ENDPOINTS: <set to the key 'data' of config map 'rook-ceph-mon-endpoints'> Optional: false
  285. ROOK_MON_SECRET: <set to the key 'mon-secret' in secret 'rook-ceph-mon'> Optional: false
  286. ROOK_CEPH_USERNAME: <set to the key 'ceph-username' in secret 'rook-ceph-mon'> Optional: false
  287. ROOK_CEPH_SECRET: <set to the key 'ceph-secret' in secret 'rook-ceph-mon'> Optional: false
  288. ROOK_CONFIG_DIR: /var/lib/rook
  289. ROOK_CEPH_CONFIG_OVERRIDE: /etc/rook/config/override.conf
  290. ROOK_FSID: <set to the key 'fsid' in secret 'rook-ceph-mon'> Optional: false
  291. NODE_NAME: (v1:spec.nodeName)
  292. ROOK_CRUSHMAP_ROOT: default
  293. ROOK_CRUSHMAP_HOSTNAME:
  294. CEPH_VOLUME_DEBUG: 1
  295. CEPH_VOLUME_SKIP_RESTORECON: 1
  296. DM_DISABLE_UDEV: 1
  297. TINI_SUBREAPER:
  298. CONTAINER_IMAGE: ceph/daemon-base:latest-pacific
  299. POD_NAME: rook-ceph-osd-0-75765667f8-pl4c9 (v1:metadata.name)
  300. POD_MEMORY_LIMIT: 5368709120 (limits.memory)
  301. POD_MEMORY_REQUEST: 5368709120 (requests.memory)
  302. POD_CPU_LIMIT: 2 (limits.cpu)
  303. POD_CPU_REQUEST: 2 (requests.cpu)
  304. ROOK_OSD_UUID: 1fa42919-fc32-4ab5-a06a-e2c93b33c64a
  305. ROOK_OSD_ID: 0
  306. ROOK_CEPH_MON_HOST: <set to the key 'mon_host' in secret 'rook-ceph-config'> Optional: false
  307. CEPH_ARGS: -m $(ROOK_CEPH_MON_HOST)
  308. ROOK_BLOCK_PATH: /mnt/example-deviceset-1-data-0plqzb
  309. ROOK_CV_MODE: raw
  310. ROOK_OSD_PVC_SIZE: 100Gi
  311. ROOK_PVC_BACKED_OSD: true
  312. Mounts:
  313. /etc/ceph from rook-config-override (ro)
  314. /run/udev from run-udev (rw)
  315. /var/lib/ceph/crash from rook-ceph-crash (rw)
  316. /var/lib/ceph/osd/ceph-0 from example-deviceset-1-data-0plqzb-bridge (rw,path="ceph-0")
  317. /var/lib/rook from rook-data (rw)
  318. /var/log/ceph from rook-ceph-log (rw)
  319. /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
  320. log-collector:
  321. Container ID: cri-o://fbc480fd8763a76417f618ca6cad0f822dcba2efa801fdcc5696a136ff18672b
  322. Image: ceph/daemon-base:latest-pacific
  323. Image ID: docker.io/ceph/daemon-base@sha256:604acd504b9a239614f6bc4f2c56f943c64ab46339added7a3296f50cc4ccec6
  324. Port: <none>
  325. Host Port: <none>
  326. Command:
  327. /bin/bash
  328. -c
  329.  
  330. set -xe
  331.  
  332. CEPH_CLIENT_ID=ceph-osd.0
  333. PERIODICITY=24h
  334. LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph
  335.  
  336. if [ -z "$PERIODICITY" ]; then
  337. PERIODICITY=24h
  338. fi
  339.  
  340. # edit the logrotate file to only rotate a specific daemon log
  341. # otherwise we will logrotate log files without reloading certain daemons
  342. # this might happen when multiple daemons run on the same machine
  343. sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE"
  344.  
  345. while true; do
  346. sleep "$PERIODICITY"
  347. echo "starting log rotation"
  348. logrotate --verbose --force "$LOG_ROTATE_CEPH_FILE"
  349. echo "I am going to sleep now, see you in $PERIODICITY"
  350. done
  351.  
  352. State: Running
  353. Started: Mon, 07 Jun 2021 15:44:13 +0530
  354. Ready: True
  355. Restart Count: 0
  356. Environment: <none>
  357. Mounts:
  358. /etc/ceph from rook-config-override (ro)
  359. /var/lib/ceph/crash from rook-ceph-crash (rw)
  360. /var/log/ceph from rook-ceph-log (rw)
  361. /var/run/secrets/kubernetes.io/serviceaccount from rook-ceph-osd-token-qljzb (ro)
  362. Conditions:
  363. Type Status
  364. Initialized True
  365. Ready False
  366. ContainersReady False
  367. PodScheduled True
  368. Volumes:
  369. rook-data:
  370. Type: EmptyDir (a temporary directory that shares a pod's lifetime)
  371. Medium:
  372. SizeLimit: <unset>
  373. rook-config-override:
  374. Type: Projected (a volume that contains injected data from multiple sources)
  375. ConfigMapName: rook-config-override
  376. ConfigMapOptional: <nil>
  377. rook-ceph-log:
  378. Type: HostPath (bare host directory volume)
  379. Path: /var/lib/rook/openshift-storage/log
  380. HostPathType:
  381. rook-ceph-crash:
  382. Type: HostPath (bare host directory volume)
  383. Path: /var/lib/rook/openshift-storage/crash
  384. HostPathType:
  385. example-deviceset-1-data-0plqzb:
  386. Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace)
  387. ClaimName: example-deviceset-1-data-0plqzb
  388. ReadOnly: false
  389. example-deviceset-1-data-0plqzb-bridge:
  390. Type: HostPath (bare host directory volume)
  391. Path: /var/lib/rook/openshift-storage/example-deviceset-1-data-0plqzb
  392. HostPathType: DirectoryOrCreate
  393. run-udev:
  394. Type: HostPath (bare host directory volume)
  395. Path: /run/udev
  396. HostPathType:
  397. rook-ceph-osd-token-qljzb:
  398. Type: Secret (a volume populated by a Secret)
  399. SecretName: rook-ceph-osd-token-qljzb
  400. Optional: false
  401. QoS Class: Burstable
  402. Node-Selectors: kubernetes.io/hostname=rhocs-bm8.lab.eng.blr.redhat.com
  403. Tolerations: node.kubernetes.io/memory-pressure:NoSchedule op=Exists
  404. node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
  405. node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
  406. node.ocs.openshift.io/storage=true:NoSchedule
  407. Events:
  408. Type Reason Age From Message
  409. ---- ------ ---- ---- -------
  410. Normal Scheduled 108m default-scheduler Successfully assigned openshift-storage/rook-ceph-osd-0-75765667f8-pl4c9 to rhocs-bm8.lab.eng.blr.redhat.com
  411. Normal SuccessfulMountVolume 108m kubelet MapVolume.MapPodDevice succeeded for volume "local-pv-6b183b4d" globalMapPath "/var/lib/kubelet/plugins/kubernetes.io~local-volume/volumeDevices/local-pv-6b183b4d"
  412. Normal SuccessfulMountVolume 108m kubelet MapVolume.MapPodDevice succeeded for volume "local-pv-6b183b4d" volumeMapPath "/var/lib/kubelet/pods/88e797db-e8bc-42fa-a5cd-2e1f9ff5ac55/volumeDevices/kubernetes.io~local-volume"
  413. Normal AddedInterface 108m multus Add eth0 [10.130.0.251/23]
  414. Normal AddedInterface 108m multus Add net1 [2620:52:0:463e:22e8:b8fb:89ae:3813/117] from openshift-storage/ipv6
  415. Normal AddedInterface 108m multus Add net2 [2620:52:0:463e:22e8:b8fb:89ae:3814/117] from openshift-storage/ipv6
  416. Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
  417. Normal Created 108m kubelet Created container blkdevmapper
  418. Normal Started 108m kubelet Started container blkdevmapper
  419. Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
  420. Normal Created 108m kubelet Created container activate
  421. Normal Started 108m kubelet Started container activate
  422. Normal Started 108m kubelet Started container expand-bluefs
  423. Normal Created 108m kubelet Created container expand-bluefs
  424. Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
  425. Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
  426. Normal Created 108m kubelet Created container chown-container-data-dir
  427. Normal Started 108m kubelet Started container chown-container-data-dir
  428. Normal Pulled 108m kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
  429. Normal Pulled 108m (x2 over 108m) kubelet Container image "ceph/daemon-base:latest-pacific" already present on machine
  430. Normal Created 108m (x2 over 108m) kubelet Created container osd
  431. Normal Started 108m (x2 over 108m) kubelet Started container osd
  432. Normal Created 108m kubelet Created container log-collector
  433. Normal Started 108m kubelet Started container log-collector
  434. Warning BackOff 3m43s (x503 over 108m) kubelet Back-off restarting failed container
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement