Advertisement
rzerres

ceph-pg_unfound

Apr 11th, 2013
26
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 14.15 KB | None | 0 0
  1. # ceph -s
  2. health HEALTH_WARN 1 pgs recovering; 50 pgs stale; 50 pgs stuck stale; 1 pgs stuck unclean; recovery 7714/2278940 degraded (0.338%); 3857/1139470 unfound (0.338%)
  3. monmap e10: 3 mons at {a=172.16.0.3:6789/0,b=172.16.0.4:6789/0,c=172.16.0.2:6789/0}, election epoch 350, quorum 0,1,2 a,b,c
  4. osdmap e4832: 8 osds: 8 up, 8 in
  5. pgmap v2124426: 1192 pgs: 1140 active+clean, 50 stale+active+clean, 1 active+clean+scrubbing+deep, 1 active+recovering; 4427 GB data, 8834 GB used, 9660 GB / 18588 GB avail; 7714/2278940 degraded (0.338%); 3857/1139470 unfound (0.338%)
  6. mdsmap e1378: 1/1/1 up {0=0=up:active}
  7.  
  8. # ceph health detail
  9. HEALTH_WARN 1 pgs recovering; 50 pgs stale; 50 pgs stuck stale; 1 pgs stuck unclean; recovery 7714/2278940 degraded (0.338%); 3857/1139470 unfound (0.338%)
  10. pg 11.2 is stuck unclean since forever, current state active+recovering, last acting [6,8]
  11. pg 2.65 is stuck stale for 2067254.371209, current state stale+active+clean, last acting [7,0]
  12. pg 0.67 is stuck stale for 2067254.371228, current state stale+active+clean, last acting [7,0]
  13. pg 1.66 is stuck stale for 2067254.371243, current state stale+active+clean, last acting [7,0]
  14. pg 7.60 is stuck stale for 2067254.371256, current state stale+active+clean, last acting [7,0]
  15. pg 2.64 is stuck stale for 2067254.371258, current state stale+active+clean, last acting [0,8]
  16. pg 0.66 is stuck stale for 2067254.371270, current state stale+active+clean, last acting [0,8]
  17. pg 1.65 is stuck stale for 2067254.371274, current state stale+active+clean, last acting [0,8]
  18. pg 1.60 is stuck stale for 2067254.371273, current state stale+active+clean, last acting [1,7]
  19. pg 0.61 is stuck stale for 2067254.371283, current state stale+active+clean, last acting [1,7]
  20. pg 7.5a is stuck stale for 2067254.371283, current state stale+active+clean, last acting [1,7]
  21. pg 2.5f is stuck stale for 2067254.371291, current state stale+active+clean, last acting [1,7]
  22. pg 7.5f is stuck stale for 2067254.371288, current state stale+active+clean, last acting [0,8]
  23. pg 0.42 is stuck stale for 2067254.371221, current state stale+active+clean, last acting [7,1]
  24. pg 2.40 is stuck stale for 2067254.371231, current state stale+active+clean, last acting [7,1]
  25. pg 1.41 is stuck stale for 2067254.371234, current state stale+active+clean, last acting [7,1]
  26. pg 7.3b is stuck stale for 2067254.371232, current state stale+active+clean, last acting [7,1]
  27. pg 1.2c is stuck stale for 2067254.371192, current state stale+active+clean, last acting [1,8]
  28. pg 0.2d is stuck stale for 2067254.371202, current state stale+active+clean, last acting [1,8]
  29. pg 2.2b is stuck stale for 2067254.371199, current state stale+active+clean, last acting [1,8]
  30. pg 0.28 is stuck stale for 2067254.371206, current state stale+active+clean, last acting [0,8]
  31. pg 7.21 is stuck stale for 2067254.371211, current state stale+active+clean, last acting [0,8]
  32. pg 1.27 is stuck stale for 2067254.371221, current state stale+active+clean, last acting [0,8]
  33. pg 2.26 is stuck stale for 2067254.371220, current state stale+active+clean, last acting [0,8]
  34. pg 7.26 is stuck stale for 2067254.371224, current state stale+active+clean, last acting [1,8]
  35. pg 7.10 is stuck stale for 2067254.371200, current state stale+active+clean, last acting [7,0]
  36. pg 1.16 is stuck stale for 2067254.371208, current state stale+active+clean, last acting [7,0]
  37. pg 0.17 is stuck stale for 2067254.371218, current state stale+active+clean, last acting [7,0]
  38. pg 2.15 is stuck stale for 2067254.371226, current state stale+active+clean, last acting [7,0]
  39. pg 7.bb is stuck stale for 2067254.371886, current state stale+active+clean, last acting [7,1]
  40. pg 7.a1 is stuck stale for 2067254.371823, current state stale+active+clean, last acting [0,8]
  41. pg 7.a6 is stuck stale for 2067254.371817, current state stale+active+clean, last acting [1,8]
  42. pg 0.9e is stuck stale for 2067254.371814, current state stale+active+clean, last acting [7,1]
  43. pg 2.9c is stuck stale for 2067254.371824, current state stale+active+clean, last acting [7,1]
  44. pg 1.9d is stuck stale for 2067254.371830, current state stale+active+clean, last acting [7,1]
  45. pg 2.98 is stuck stale for 2067254.371832, current state stale+active+clean, last acting [7,0]
  46. pg 0.9a is stuck stale for 2067254.371844, current state stale+active+clean, last acting [7,0]
  47. pg 1.99 is stuck stale for 2067254.371847, current state stale+active+clean, last acting [7,0]
  48. pg 2.95 is stuck stale for 2067254.371848, current state stale+active+clean, last acting [1,7]
  49. pg 0.97 is stuck stale for 2067254.371860, current state stale+active+clean, last acting [1,7]
  50. pg 1.96 is stuck stale for 2067254.371870, current state stale+active+clean, last acting [1,7]
  51. pg 7.90 is stuck stale for 2067254.371885, current state stale+active+clean, last acting [1,7]
  52. pg 7.93 is stuck stale for 2067254.371879, current state stale+active+clean, last acting [7,0]
  53. pg 1.88 is stuck stale for 2067254.371855, current state stale+active+clean, last acting [7,1]
  54. pg 0.89 is stuck stale for 2067254.371865, current state stale+active+clean, last acting [7,1]
  55. pg 7.82 is stuck stale for 2067254.371861, current state stale+active+clean, last acting [7,1]
  56. pg 2.87 is stuck stale for 2067254.371868, current state stale+active+clean, last acting [7,1]
  57. pg 2.80 is stuck stale for 2067254.371867, current state stale+active+clean, last acting [1,8]
  58. pg 0.82 is stuck stale for 2067254.371879, current state stale+active+clean, last acting [1,8]
  59. pg 1.81 is stuck stale for 2067254.371875, current state stale+active+clean, last acting [1,8]
  60. pg 7.7b is stuck stale for 2067254.371870, current state stale+active+clean, last acting [1,8]
  61. pg 11.2 is active+recovering, acting [6,8], 3857 unfound
  62. recovery 7714/2278940 degraded (0.338%); 3857/1139470 unfound (0.338%)
  63.  
  64. # ceph pg 1.81 query
  65. i don't have pgid 1.81
  66.  
  67.  
  68. # ceph pg 11.2 query
  69.  
  70. { "state": "active+recovering",
  71. "epoch": 4832,
  72. "up": [
  73. 6,
  74. 8],
  75. "acting": [
  76. 6,
  77. 8],
  78. "info": { "pgid": "11.2",
  79. "last_update": "4084'7092",
  80. "last_complete": "2757'407",
  81. "log_tail": "0'0",
  82. "last_backfill": "MAX",
  83. "purged_snaps": "[]",
  84. "history": { "epoch_created": 2726,
  85. "last_epoch_started": 4827,
  86. "last_epoch_clean": 3294,
  87. "last_epoch_split": 0,
  88. "same_up_since": 4826,
  89. "same_interval_since": 4826,
  90. "same_primary_since": 4813,
  91. "last_scrub": "0'0",
  92. "last_scrub_stamp": "2013-03-04 15:30:15.335201",
  93. "last_deep_scrub": "0'0",
  94. "last_deep_scrub_stamp": "2013-03-04 15:30:15.335201",
  95. "last_clean_scrub_stamp": "2013-03-04 15:30:15.335201"},
  96. "stats": { "version": "4084'7092",
  97. "reported": "4813'3473",
  98. "state": "active+recovering",
  99. "last_fresh": "2013-04-11 14:45:19.262633",
  100. "last_change": "2013-04-05 12:54:57.430937",
  101. "last_active": "2013-04-11 14:45:19.262633",
  102. "last_clean": "0.000000",
  103. "last_unstale": "2013-04-11 14:45:19.262633",
  104. "mapping_epoch": 4820,
  105. "log_start": "0'0",
  106. "ondisk_log_start": "0'0",
  107. "created": 2726,
  108. "last_epoch_clean": 2726,
  109. "parent": "0.0",
  110. "parent_split_bits": 0,
  111. "last_scrub": "0'0",
  112. "last_scrub_stamp": "2013-03-04 15:30:15.335201",
  113. "last_deep_scrub": "0'0",
  114. "last_deep_scrub_stamp": "2013-03-04 15:30:15.335201",
  115. "last_clean_scrub_stamp": "2013-03-04 15:30:15.335201",
  116. "log_size": 0,
  117. "ondisk_log_size": 0,
  118. "stats_invalid": "0",
  119. "stat_sum": { "num_bytes": 17382162544,
  120. "num_objects": 4146,
  121. "num_object_clones": 0,
  122. "num_object_copies": 0,
  123. "num_objects_missing_on_primary": 0,
  124. "num_objects_degraded": 0,
  125. "num_objects_unfound": 0,
  126. "num_read": 0,
  127. "num_read_kb": 0,
  128. "num_write": 7078,
  129. "num_write_kb": 18952638,
  130. "num_scrub_errors": 0,
  131. "num_objects_recovered": 8237,
  132. "num_bytes_recovered": 34517590688,
  133. "num_keys_recovered": 0},
  134. "stat_cat_sum": {},
  135. "up": [
  136. 6,
  137. 8],
  138. "acting": [
  139. 6,
  140. 8]},
  141. "empty": 0,
  142. "dne": 0,
  143. "incomplete": 0,
  144. "last_epoch_started": 4827},
  145. "recovery_state": [
  146. { "name": "Started\/Primary\/Active",
  147. "enter_time": "2013-04-05 12:54:57.241401",
  148. "might_have_unfound": [
  149. { "osd": 7,
  150. "status": "querying"},
  151. { "osd": 8,
  152. "status": "already probed"}],
  153. "recovery_progress": { "backfill_target": -1,
  154. "waiting_on_backfill": 0,
  155. "backfill_pos": "0\/\/0\/\/-1",
  156. "backfill_info": { "begin": "0\/\/0\/\/-1",
  157. "end": "0\/\/0\/\/-1",
  158. "objects": []},
  159. "peer_backfill_info": { "begin": "0\/\/0\/\/-1",
  160. "end": "0\/\/0\/\/-1",
  161. "objects": []},
  162. "backfills_in_flight": [],
  163. "pull_from_peer": [],
  164. "pushing": []},
  165. "scrub": { "scrubber.epoch_start": "0",
  166. "scrubber.active": 0,
  167. "scrubber.block_writes": 0,
  168. "scrubber.finalizing": 0,
  169. "scrubber.waiting_on": 0,
  170. "scrubber.waiting_on_whom": []}},
  171. { "name": "Started",
  172. "enter_time": "2013-04-05 12:54:56.037918"}]}
  173.  
  174. # ceph osd tree
  175.  
  176. # id weight type name up/down reweight
  177. -1 16.8 room server-raum-keller
  178. -2 16.8 rack rack-daywalker
  179. -4 14.8 storage daywalker-data
  180. -6 7.4 host dwssrv1
  181. 0 3.7 osd.0 up 1
  182. 1 3.7 osd.1 up 1
  183. -7 7.4 host dwssrv2
  184. 3 3.7 osd.3 up 1
  185. 4 3.7 osd.4 up 1
  186. -5 2 storage daywalker-archive
  187. -8 2 host dwssrv1-archive
  188. 7 1 osd.7 up 1
  189. 8 1 osd.8 up 1
  190. -9 2 host dwssrv2-archive
  191. 5 1 osd.5 up 1
  192. 6 1 osd.6 up 1
  193.  
  194. # rbd info dws-archive/vdi-dws-archive
  195. rbd image 'vdi-dws-archive':
  196. size 1000 GB in 256000 objects
  197. order 22 (4096 KB objects)
  198. block_name_prefix: rb.0.e6d2.238e1f29
  199. format: 1
  200.  
  201.  
  202. ##
  203. # the crushmap
  204. #
  205. # begin crush map
  206. tunable choose_local_tries 0
  207. tunable choose_local_fallback_tries 0
  208. tunable choose_total_tries 50
  209. tunable chooseleaf_descend_once 1
  210.  
  211. # devices
  212. device 0 osd.0
  213. device 1 osd.1
  214. device 2 device2
  215. device 3 osd.3
  216. device 4 osd.4
  217. device 5 osd.5
  218. device 6 osd.6
  219. device 7 osd.7
  220. device 8 osd.8
  221.  
  222. # types
  223. type 0 osd
  224. type 1 host
  225. type 2 storage
  226. type 3 rack
  227. type 4 room
  228.  
  229. # buckets
  230. host dwssrv1 {
  231. id -6 # do not change unnecessarily
  232. # weight 7.400
  233. alg straw
  234. hash 0 # rjenkins1
  235. item osd.0 weight 3.700
  236. item osd.1 weight 3.700
  237. }
  238. host dwssrv2 {
  239. id -7 # do not change unnecessarily
  240. # weight 7.400
  241. alg straw
  242. hash 0 # rjenkins1
  243. item osd.3 weight 3.700
  244. item osd.4 weight 3.700
  245. }
  246. storage daywalker-data {
  247. id -4 # do not change unnecessarily
  248. # weight 14.800
  249. alg straw
  250. hash 0 # rjenkins1
  251. item dwssrv1 weight 7.400
  252. item dwssrv2 weight 7.400
  253. }
  254. host dwssrv1-archive {
  255. id -8 # do not change unnecessarily
  256. # weight 2.000
  257. alg straw
  258. hash 0 # rjenkins1
  259. item osd.7 weight 1.000
  260. item osd.8 weight 1.000
  261. }
  262. host dwssrv2-archive {
  263. id -9 # do not change unnecessarily
  264. # weight 2.000
  265. alg straw
  266. hash 0 # rjenkins1
  267. item osd.5 weight 1.000
  268. item osd.6 weight 1.000
  269. }
  270. storage daywalker-archive {
  271. id -5 # do not change unnecessarily
  272. # weight 4.000
  273. alg straw
  274. hash 0 # rjenkins1
  275. item dwssrv1-archive weight 2.000
  276. item dwssrv2-archive weight 2.000
  277. }
  278. rack rack-daywalker {
  279. id -2 # do not change unnecessarily
  280. # weight 16.800
  281. alg straw
  282. hash 0 # rjenkins1
  283. item daywalker-data weight 14.800
  284. item daywalker-archive weight 2.000
  285. }
  286. room server-raum-keller {
  287. id -1 # do not change unnecessarily
  288. # weight 16.800
  289. alg straw
  290. hash 0 # rjenkins1
  291. item rack-daywalker weight 16.800
  292. }
  293.  
  294. # rules
  295. rule data {
  296. ruleset 0
  297. type replicated
  298. min_size 1
  299. max_size 10
  300. step take server-raum-keller
  301. step chooseleaf firstn 0 type host
  302. step emit
  303. }
  304. rule metadata {
  305. ruleset 1
  306. type replicated
  307. min_size 1
  308. max_size 10
  309. step take server-raum-keller
  310. step chooseleaf firstn 0 type host
  311. step emit
  312. }
  313. rule rbd {
  314. ruleset 2
  315. type replicated
  316. min_size 1
  317. max_size 10
  318. step take server-raum-keller
  319. step chooseleaf firstn 0 type host
  320. step emit
  321. }
  322. rule dws-vdi {
  323. ruleset 3
  324. type replicated
  325. min_size 1
  326. max_size 10
  327. step take daywalker-data
  328. step chooseleaf firstn 0 type host
  329. step emit
  330. }
  331. rule dws-archive {
  332. ruleset 4
  333. type replicated
  334. min_size 1
  335. max_size 10
  336. step take daywalker-archive
  337. step chooseleaf firstn 0 type host
  338. step emit
  339. }
  340. rule dws-test {
  341. ruleset 5
  342. type replicated
  343. min_size 1
  344. max_size 10
  345. step take rack-daywalker
  346. step chooseleaf firstn 0 type host
  347. step emit
  348. }
  349.  
  350. # end crush map
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement