Advertisement
Proff_Ust

Untitled

Jul 6th, 2018
96
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 9.25 KB | None | 0 0
  1. ceph,collection=WBThrottle,host=ceph-osd-mon01,id=0,type=osd
  2. ceph,collection=WBThrottle,host=ceph-osd-mon01,id=1,type=osd
  3. ceph,collection=WBThrottle,host=ceph-osd-mon01,id=2,type=osd
  4. ceph,collection=cluster,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  5. ceph,collection=filestore,host=ceph-osd-mon01,id=0,type=osd
  6. ceph,collection=filestore,host=ceph-osd-mon01,id=1,type=osd
  7. ceph,collection=filestore,host=ceph-osd-mon01,id=2,type=osd
  8. ceph,collection=finisher-JournalObjectStore,host=ceph-osd-mon01,id=0,type=osd
  9. ceph,collection=finisher-JournalObjectStore,host=ceph-osd-mon01,id=1,type=osd
  10. ceph,collection=finisher-JournalObjectStore,host=ceph-osd-mon01,id=2,type=osd
  11. ceph,collection=finisher-filestore-apply-0,host=ceph-osd-mon01,id=0,type=osd
  12. ceph,collection=finisher-filestore-apply-0,host=ceph-osd-mon01,id=1,type=osd
  13. ceph,collection=finisher-filestore-apply-0,host=ceph-osd-mon01,id=2,type=osd
  14. ceph,collection=finisher-filestore-ondisk-0,host=ceph-osd-mon01,id=0,type=osd
  15. ceph,collection=finisher-filestore-ondisk-0,host=ceph-osd-mon01,id=1,type=osd
  16. ceph,collection=finisher-filestore-ondisk-0,host=ceph-osd-mon01,id=2,type=osd
  17. ceph,collection=finisher-monstore,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  18. ceph,collection=leveldb,host=ceph-osd-mon01,id=0,type=osd
  19. ceph,collection=leveldb,host=ceph-osd-mon01,id=1,type=osd
  20. ceph,collection=leveldb,host=ceph-osd-mon01,id=2,type=osd
  21. ceph,collection=leveldb,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  22. ceph,collection=mon,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  23. ceph,collection=mutex-FileJournal::completions_lock,host=ceph-osd-mon01,id=0,type=osd
  24. ceph,collection=mutex-FileJournal::completions_lock,host=ceph-osd-mon01,id=1,type=osd
  25. ceph,collection=mutex-FileJournal::completions_lock,host=ceph-osd-mon01,id=2,type=osd
  26. ceph,collection=mutex-FileJournal::finisher_lock,host=ceph-osd-mon01,id=0,type=osd
  27. ceph,collection=mutex-FileJournal::finisher_lock,host=ceph-osd-mon01,id=1,type=osd
  28. ceph,collection=mutex-FileJournal::finisher_lock,host=ceph-osd-mon01,id=2,type=osd
  29. ceph,collection=mutex-FileJournal::write_lock,host=ceph-osd-mon01,id=0,type=osd
  30. ceph,collection=mutex-FileJournal::write_lock,host=ceph-osd-mon01,id=1,type=osd
  31. ceph,collection=mutex-FileJournal::write_lock,host=ceph-osd-mon01,id=2,type=osd
  32. ceph,collection=mutex-FileJournal::writeq_lock,host=ceph-osd-mon01,id=0,type=osd
  33. ceph,collection=mutex-FileJournal::writeq_lock,host=ceph-osd-mon01,id=1,type=osd
  34. ceph,collection=mutex-FileJournal::writeq_lock,host=ceph-osd-mon01,id=2,type=osd
  35. ceph,collection=mutex-JOS::ApplyManager::apply_lock,host=ceph-osd-mon01,id=0,type=osd
  36. ceph,collection=mutex-JOS::ApplyManager::apply_lock,host=ceph-osd-mon01,id=1,type=osd
  37. ceph,collection=mutex-JOS::ApplyManager::apply_lock,host=ceph-osd-mon01,id=2,type=osd
  38. ceph,collection=mutex-JOS::ApplyManager::com_lock,host=ceph-osd-mon01,id=0,type=osd
  39. ceph,collection=mutex-JOS::ApplyManager::com_lock,host=ceph-osd-mon01,id=1,type=osd
  40. ceph,collection=mutex-JOS::ApplyManager::com_lock,host=ceph-osd-mon01,id=2,type=osd
  41. ceph,collection=mutex-JOS::SubmitManager::lock,host=ceph-osd-mon01,id=0,type=osd
  42. ceph,collection=mutex-JOS::SubmitManager::lock,host=ceph-osd-mon01,id=1,type=osd
  43. ceph,collection=mutex-JOS::SubmitManager::lock,host=ceph-osd-mon01,id=2,type=osd
  44. ceph,collection=mutex-OSD:ShardedOpWQ:.0,host=ceph-osd-mon01,id=0,type=osd
  45. ceph,collection=mutex-OSD:ShardedOpWQ:.0,host=ceph-osd-mon01,id=1,type=osd
  46. ceph,collection=mutex-OSD:ShardedOpWQ:.0,host=ceph-osd-mon01,id=2,type=osd
  47. ceph,collection=mutex-OSD:ShardedOpWQ:.1,host=ceph-osd-mon01,id=0,type=osd
  48. ceph,collection=mutex-OSD:ShardedOpWQ:.1,host=ceph-osd-mon01,id=1,type=osd
  49. ceph,collection=mutex-OSD:ShardedOpWQ:.1,host=ceph-osd-mon01,id=2,type=osd
  50. ceph,collection=mutex-OSD:ShardedOpWQ:.2,host=ceph-osd-mon01,id=0,type=osd
  51. ceph,collection=mutex-OSD:ShardedOpWQ:.2,host=ceph-osd-mon01,id=1,type=osd
  52. ceph,collection=mutex-OSD:ShardedOpWQ:.2,host=ceph-osd-mon01,id=2,type=osd
  53. ceph,collection=mutex-OSD:ShardedOpWQ:.3,host=ceph-osd-mon01,id=0,type=osd
  54. ceph,collection=mutex-OSD:ShardedOpWQ:.3,host=ceph-osd-mon01,id=1,type=osd
  55. ceph,collection=mutex-OSD:ShardedOpWQ:.3,host=ceph-osd-mon01,id=2,type=osd
  56. ceph,collection=mutex-OSD:ShardedOpWQ:.4,host=ceph-osd-mon01,id=0,type=osd
  57. ceph,collection=mutex-OSD:ShardedOpWQ:.4,host=ceph-osd-mon01,id=1,type=osd
  58. ceph,collection=mutex-OSD:ShardedOpWQ:.4,host=ceph-osd-mon01,id=2,type=osd
  59. ceph,collection=mutex-OSD:ShardedOpWQ:order:.0,host=ceph-osd-mon01,id=0,type=osd
  60. ceph,collection=mutex-OSD:ShardedOpWQ:order:.0,host=ceph-osd-mon01,id=1,type=osd
  61. ceph,collection=mutex-OSD:ShardedOpWQ:order:.0,host=ceph-osd-mon01,id=2,type=osd
  62. ceph,collection=mutex-OSD:ShardedOpWQ:order:.1,host=ceph-osd-mon01,id=0,type=osd
  63. ceph,collection=mutex-OSD:ShardedOpWQ:order:.1,host=ceph-osd-mon01,id=1,type=osd
  64. ceph,collection=mutex-OSD:ShardedOpWQ:order:.1,host=ceph-osd-mon01,id=2,type=osd
  65. ceph,collection=mutex-OSD:ShardedOpWQ:order:.2,host=ceph-osd-mon01,id=0,type=osd
  66. ceph,collection=mutex-OSD:ShardedOpWQ:order:.2,host=ceph-osd-mon01,id=1,type=osd
  67. ceph,collection=mutex-OSD:ShardedOpWQ:order:.2,host=ceph-osd-mon01,id=2,type=osd
  68. ceph,collection=mutex-OSD:ShardedOpWQ:order:.3,host=ceph-osd-mon01,id=0,type=osd
  69. ceph,collection=mutex-OSD:ShardedOpWQ:order:.3,host=ceph-osd-mon01,id=1,type=osd
  70. ceph,collection=mutex-OSD:ShardedOpWQ:order:.3,host=ceph-osd-mon01,id=2,type=osd
  71. ceph,collection=mutex-OSD:ShardedOpWQ:order:.4,host=ceph-osd-mon01,id=0,type=osd
  72. ceph,collection=mutex-OSD:ShardedOpWQ:order:.4,host=ceph-osd-mon01,id=1,type=osd
  73. ceph,collection=mutex-OSD:ShardedOpWQ:order:.4,host=ceph-osd-mon01,id=2,type=osd
  74. ceph,collection=mutex-WBThrottle::lock,host=ceph-osd-mon01,id=0,type=osd
  75. ceph,collection=mutex-WBThrottle::lock,host=ceph-osd-mon01,id=1,type=osd
  76. ceph,collection=mutex-WBThrottle::lock,host=ceph-osd-mon01,id=2,type=osd
  77. ceph,collection=objecter,host=ceph-osd-mon01,id=0,type=osd
  78. ceph,collection=objecter,host=ceph-osd-mon01,id=1,type=osd
  79. ceph,collection=objecter,host=ceph-osd-mon01,id=2,type=osd
  80. ceph,collection=osd,host=ceph-osd-mon01,id=0,type=osd
  81. ceph,collection=osd,host=ceph-osd-mon01,id=1,type=osd
  82. ceph,collection=osd,host=ceph-osd-mon01,id=2,type=osd
  83. ceph,collection=paxos,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  84. ceph,collection=recoverystate_perf,host=ceph-osd-mon01,id=0,type=osd
  85. ceph,collection=recoverystate_perf,host=ceph-osd-mon01,id=1,type=osd
  86. ceph,collection=recoverystate_perf,host=ceph-osd-mon01,id=2,type=osd
  87. ceph,collection=throttle-mon_client_bytes,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  88. ceph,collection=throttle-mon_daemon_bytes,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  89. ceph,collection=throttle-msgr_dispatch_throttler-client,host=ceph-osd-mon01,id=0,type=osd
  90. ceph,collection=throttle-msgr_dispatch_throttler-client,host=ceph-osd-mon01,id=1,type=osd
  91. ceph,collection=throttle-msgr_dispatch_throttler-client,host=ceph-osd-mon01,id=2,type=osd
  92. ceph,collection=throttle-msgr_dispatch_throttler-cluster,host=ceph-osd-mon01,id=0,type=osd
  93. ceph,collection=throttle-msgr_dispatch_throttler-cluster,host=ceph-osd-mon01,id=1,type=osd
  94. ceph,collection=throttle-msgr_dispatch_throttler-cluster,host=ceph-osd-mon01,id=2,type=osd
  95. ceph,collection=throttle-msgr_dispatch_throttler-hb_back_server,host=ceph-osd-mon01,id=0,type=osd
  96. ceph,collection=throttle-msgr_dispatch_throttler-hb_back_server,host=ceph-osd-mon01,id=1,type=osd
  97. ceph,collection=throttle-msgr_dispatch_throttler-hb_back_server,host=ceph-osd-mon01,id=2,type=osd
  98. ceph,collection=throttle-msgr_dispatch_throttler-hb_front_server,host=ceph-osd-mon01,id=0,type=osd
  99. ceph,collection=throttle-msgr_dispatch_throttler-hb_front_server,host=ceph-osd-mon01,id=1,type=osd
  100. ceph,collection=throttle-msgr_dispatch_throttler-hb_front_server,host=ceph-osd-mon01,id=2,type=osd
  101. ceph,collection=throttle-msgr_dispatch_throttler-hbclient,host=ceph-osd-mon01,id=0,type=osd
  102. ceph,collection=throttle-msgr_dispatch_throttler-hbclient,host=ceph-osd-mon01,id=1,type=osd
  103. ceph,collection=throttle-msgr_dispatch_throttler-hbclient,host=ceph-osd-mon01,id=2,type=osd
  104. ceph,collection=throttle-msgr_dispatch_throttler-mon,host=ceph-osd-mon01,id=ceph-osd-mon01,type=monitor
  105. ceph,collection=throttle-msgr_dispatch_throttler-ms_objecter,host=ceph-osd-mon01,id=0,type=osd
  106. ceph,collection=throttle-msgr_dispatch_throttler-ms_objecter,host=ceph-osd-mon01,id=1,type=osd
  107. ceph,collection=throttle-msgr_dispatch_throttler-ms_objecter,host=ceph-osd-mon01,id=2,type=osd
  108. ceph,collection=throttle-objecter_bytes,host=ceph-osd-mon01,id=0,type=osd
  109. ceph,collection=throttle-objecter_bytes,host=ceph-osd-mon01,id=1,type=osd
  110. ceph,collection=throttle-objecter_bytes,host=ceph-osd-mon01,id=2,type=osd
  111. ceph,collection=throttle-objecter_ops,host=ceph-osd-mon01,id=0,type=osd
  112. ceph,collection=throttle-objecter_ops,host=ceph-osd-mon01,id=1,type=osd
  113. ceph,collection=throttle-objecter_ops,host=ceph-osd-mon01,id=2,type=osd
  114. ceph,collection=throttle-osd_client_bytes,host=ceph-osd-mon01,id=0,type=osd
  115. ceph,collection=throttle-osd_client_bytes,host=ceph-osd-mon01,id=1,type=osd
  116. ceph,collection=throttle-osd_client_bytes,host=ceph-osd-mon01,id=2,type=osd
  117. ceph,collection=throttle-osd_client_messages,host=ceph-osd-mon01,id=0,type=osd
  118. ceph,collection=throttle-osd_client_messages,host=ceph-osd-mon01,id=1,type=osd
  119. ceph,collection=throttle-osd_client_messages,host=ceph-osd-mon01,id=2,type=osd
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement