Advertisement
Guest User

ceph segfault

a guest
Nov 21st, 2012
16
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 24.90 KB | None | 0 0
  1. root@hotcpc9575:/# /usr/bin/ceph-mon --debug_mon 10 -d --debug_ms 3 -i a --pid-file /var/run/ceph/mon.a.pid -c /etc/ceph/ceph.conf
  2. 2012-11-21 11:07:03.409614 7f3db7706780 0 ceph version 0.54 (commit:60b84b095b1009a305d4d6a5b16f88571cbd3150), process ceph-mon, pid 15637
  3. 2012-11-21 11:07:03.409627 7f3db7706780 1 store(/var/lib/ceph/mon/ceph-a) mount
  4. starting mon.a rank 0 at 10.252.24.92:6789/0 mon_data /var/lib/ceph/mon/ceph-a fsid 1b0277a7-d8a0-46a2-ae5a-48c59d242853
  5. 2012-11-21 11:07:03.410212 7f3db7706780 1 -- 10.252.24.92:6789/0 learned my addr 10.252.24.92:6789/0
  6. 2012-11-21 11:07:03.410228 7f3db7706780 1 accepter.accepter.bind my_inst.addr is 10.252.24.92:6789/0 need_addr=0
  7. 2012-11-21 11:07:03.411991 7f3db7706780 1 -- 10.252.24.92:6789/0 messenger.start
  8. 2012-11-21 11:07:03.411996 7f3db7706780 1 accepter.accepter.start
  9. 2012-11-21 11:07:03.412109 7f3db21c7700 1 -- 10.252.24.92:6789/0 >> :/0 pipe(0x298a900 sd=8 :6789 pgs=0 cs=0 l=0).accept sd=8
  10. 2012-11-21 11:07:03.412119 7f3db20c6700 1 -- 10.252.24.92:6789/0 >> :/0 pipe(0x298ad80 sd=9 :6789 pgs=0 cs=0 l=0).accept sd=9
  11. 2012-11-21 11:07:03.412155 7f3db1fc5700 1 -- 10.252.24.92:6789/0 >> :/0 pipe(0x298ab40 sd=10 :6789 pgs=0 cs=0 l=0).accept sd=10
  12. 2012-11-21 11:07:03.412205 7f3db7706780 1 mon.a@-1(probing) e1 init fsid 1b0277a7-d8a0-46a2-ae5a-48c59d242853
  13. 2012-11-21 11:07:03.412303 7f3db7706780 10 mon.a@-1(probing) e1 check_fsid cluster_uuid contains '1b0277a7-d8a0-46a2-ae5a-48c59d242853'
  14. 2012-11-21 11:07:03.412326 7f3db7706780 10 mon.a@-1(probing) e1 features compat={},rocompat={},incompat={1=initial feature set (~v.18),2=global version sequencing (v0.52)}
  15. 2012-11-21 11:07:03.412331 7f3db7706780 10 mon.a@-1(probing) e1 require_gv_onwire
  16. 2012-11-21 11:07:03.412338 7f3db7706780 10 mon.a@-1(probing) e1 has_ever_joined = 1
  17. 2012-11-21 11:07:03.412372 7f3db1ec4700 1 -- 10.252.24.92:6789/0 >> :/0 pipe(0x298b440 sd=11 :6789 pgs=0 cs=0 l=0).accept sd=11
  18. 2012-11-21 11:07:03.412535 7f3db7706780 7 mon.a@-1(probing).pg v0 update_from_paxos loading latest full pgmap v14625
  19. 2012-11-21 11:07:03.414666 7f3db7706780 10 mon.a@-1(probing).pg v14625 send_pg_creates to 0 pgs
  20. 2012-11-21 11:07:03.414676 7f3db7706780 10 mon.a@-1(probing).pg v14625 update_logger
  21. 2012-11-21 11:07:03.414708 7f3db7706780 10 mon.a@-1(probing).mds e0 update_from_paxos paxosv 38, my e 0
  22. 2012-11-21 11:07:03.414720 7f3db7706780 10 mon.a@-1(probing).mds e0 update_from_paxos got 38
  23. 2012-11-21 11:07:03.414735 7f3db7706780 4 mon.a@-1(probing).mds e38 new map
  24. 2012-11-21 11:07:03.414739 7f3db7706780 7 mon.a@-1(probing).mds e38 print_map
  25. epoch 38
  26. flags 0
  27. created 2012-11-09 15:44:52.968043
  28. modified 2012-11-21 10:42:41.659113
  29. tableserver 0
  30. root 0
  31. session_timeout 60
  32. session_autoclose 300
  33. last_failure 35
  34. last_failure_osd_epoch 42
  35. compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object}
  36. max_mds 1
  37. in 0
  38. up {0=4498}
  39. failed
  40. stopped
  41. data_pools [0]
  42. metadata_pool 1
  43. 4498: 10.252.24.92:6800/15350 'a' mds.0.6 up:active seq 9
  44.  
  45. 2012-11-21 11:07:03.414770 7f3db7706780 10 mon.a@-1(probing).mds e38 update_logger
  46. 2012-11-21 11:07:03.414816 7f3db7706780 7 mon.a@-1(probing).osd e0 update_from_paxos loading latest full map e47
  47. 2012-11-21 11:07:03.414909 7f3db7706780 10 mon.a@-1(probing).osd e47 send_to_waiting 47
  48. 2012-11-21 11:07:03.414917 7f3db7706780 10 mon.a@-1(probing).osd e47 update_logger
  49. 2012-11-21 11:07:03.414988 7f3db7706780 7 mon.a@-1(probing).log v16222 update_from_paxos loading summary e16222
  50. 2012-11-21 11:07:03.415053 7f3db7706780 10 mon.a@-1(probing).log v16222 check_subs
  51. 2012-11-21 11:07:03.415119 7f3db7706780 10 mon.a@-1(probing).auth v298 update_from_paxos()
  52. 2012-11-21 11:07:03.415131 7f3db7706780 7 mon.a@-1(probing).auth v298 update_from_paxos loading summary e298
  53. 2012-11-21 11:07:03.415171 7f3db7706780 10 mon.a@-1(probing).auth v298 update_from_paxos() last_allocated_id=5396 max_global_id=5396
  54. 2012-11-21 11:07:03.415358 7f3db7706780 10 mon.a@-1(probing) e1 bootstrap
  55. 2012-11-21 11:07:03.415366 7f3db7706780 10 mon.a@-1(probing) e1 unregister_cluster_logger - not registered
  56. 2012-11-21 11:07:03.415368 7f3db7706780 10 mon.a@-1(probing) e1 cancel_probe_timeout (none scheduled)
  57. 2012-11-21 11:07:03.415371 7f3db7706780 0 mon.a@-1(probing) e1 my rank is now 0 (was -1)
  58. 2012-11-21 11:07:03.415372 7f3db7706780 1 -- 10.252.24.92:6789/0 mark_down_all
  59. 2012-11-21 11:07:03.415353 7f3db11be700 1 -- 10.252.24.92:6789/0 <== osd.1 10.252.24.92:6805/15504 1 ==== auth(proto 0 26 bytes epoch 1) v1 ==== 56+0+0 (3365035062 0 0) 0x2996600 con 0x29b4a00
  60. 2012-11-21 11:07:03.415439 7f3db21c7700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:6801/15428 pipe(0x298a900 sd=8 :6789 pgs=28223489 cs=1 l=1).reader couldn't read tag, Success
  61. 2012-11-21 11:07:03.415453 7f3db1fc5700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:0/15522 pipe(0x298ab40 sd=10 :6789 pgs=760120 cs=1 l=1).reader couldn't read tag, Success
  62. 2012-11-21 11:07:03.415468 7f3db1fc5700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:0/15522 pipe(0x298ab40 sd=10 :6789 pgs=760120 cs=1 l=1).fault 0: Success
  63. 2012-11-21 11:07:03.415466 7f3db21c7700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:6801/15428 pipe(0x298a900 sd=8 :6789 pgs=28223489 cs=1 l=1).fault 0: Success
  64. 2012-11-21 11:07:03.415525 7f3db1ec4700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:6800/15350 pipe(0x298b440 sd=11 :6789 pgs=29293879 cs=1 l=1).reader couldn't read tag, Success
  65. 2012-11-21 11:07:03.415542 7f3db1ec4700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:6800/15350 pipe(0x298b440 sd=11 :6789 pgs=29293879 cs=1 l=1).fault 0: Success
  66. 2012-11-21 11:07:03.415690 7f3db20c6700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:6805/15504 pipe(0x298ad80 sd=9 :6789 pgs=28207304 cs=1 l=1).reader couldn't read tag, Success
  67. 2012-11-21 11:07:03.415707 7f3db20c6700 2 -- 10.252.24.92:6789/0 >> 10.252.24.92:6805/15504 pipe(0x298ad80 sd=9 :6789 pgs=28207304 cs=1 l=1).fault 0: Success
  68. 2012-11-21 11:07:03.415724 7f3db7706780 10 mon.a@0(probing) e1 reset
  69. 2012-11-21 11:07:03.415735 7f3db7706780 1 mon.a@0(probing) e1 win_standalone_election
  70. 2012-11-21 11:07:03.415739 7f3db7706780 10 mon.a@0(probing).elector(0) bump_epoch 0 to 1
  71. 2012-11-21 11:07:03.478498 7f3db7706780 10 mon.a@0(probing) e1 reset
  72. 2012-11-21 11:07:03.478523 7f3db7706780 10 mon.a@0(leader) e1 win_election, epoch 1 quorum is 0 features are 8388607
  73. 2012-11-21 11:07:03.478550 7f3db7706780 0 log [INF] : mon.a@0 won leader election with quorum 0
  74. 2012-11-21 11:07:03.478561 7f3db7706780 1 -- 10.252.24.92:6789/0 --> mon.0 10.252.24.92:6789/0 -- log(1 entries) v1 -- ?+0 0x2997000
  75. 2012-11-21 11:07:03.478590 7f3db7706780 10 mon.a@0(leader) e1 recovered_leader 0 pgmap ()
  76. 2012-11-21 11:07:03.478596 7f3db7706780 10 mon.a@0(leader) e1 recovered_leader 1 mdsmap (0)
  77. 2012-11-21 11:07:03.478599 7f3db7706780 10 mon.a@0(leader) e1 recovered_leader 2 osdmap (0,1)
  78. 2012-11-21 11:07:03.478602 7f3db7706780 10 mon.a@0(leader) e1 recovered_leader 3 logm (0,1,2)
  79. 2012-11-21 11:07:03.478605 7f3db7706780 10 mon.a@0(leader) e1 recovered_leader 4 monmap (0,1,2,3)
  80. 2012-11-21 11:07:03.478608 7f3db7706780 10 mon.a@0(leader) e1 recovered_leader 5 auth (0,1,2,3,4)
  81. 2012-11-21 11:07:03.478610 7f3db7706780 10 mon.a@0(leader) e1 all paxos instances recovered, going writeable
  82. 2012-11-21 11:07:03.478621 7f3db7706780 10 mon.a@0(leader).pg v14625 create_pending v 14626
  83. 2012-11-21 11:07:03.478627 7f3db7706780 10 mon.a@0(leader).pg v14625 check_osd_map already seen 47 >= 47
  84. 2012-11-21 11:07:03.478632 7f3db7706780 10 mon.a@0(leader).pg v14625 update_logger
  85. 2012-11-21 11:07:03.478660 7f3db7706780 0 log [INF] : pgmap v14625: 384 pgs: 348 active+clean, 36 active+remapped; 1000 MB data, 131 GB used, 971 GB / 1161 GB avail
  86. 2012-11-21 11:07:03.478669 7f3db7706780 1 -- 10.252.24.92:6789/0 --> mon.0 10.252.24.92:6789/0 -- log(1 entries) v1 -- ?+0 0x2996e00
  87. 2012-11-21 11:07:03.478687 7f3db7706780 10 mon.a@0(leader).mds e38 create_pending e39
  88. 2012-11-21 11:07:03.478690 7f3db7706780 10 mon.a@0(leader).mds e38 e38: 1/1/1 up {0=a=up:active}
  89. 2012-11-21 11:07:03.478698 7f3db7706780 10 mon.a@0(leader).mds e38 adding 10.252.24.92:6800/15350 mds.0.6 up:active to last_beacon
  90. 2012-11-21 11:07:03.478712 7f3db7706780 10 mon.a@0(leader).mds e38 update_logger
  91. 2012-11-21 11:07:03.478722 7f3db7706780 0 log [INF] : mdsmap e38: 1/1/1 up {0=a=up:active}
  92. 2012-11-21 11:07:03.478725 7f3db7706780 1 -- 10.252.24.92:6789/0 --> mon.0 10.252.24.92:6789/0 -- log(2 entries) v1 -- ?+0 0x2997a00
  93. 2012-11-21 11:07:03.478743 7f3db7706780 10 mon.a@0(leader).osd e47 create_pending e 48
  94. 2012-11-21 11:07:03.478747 7f3db7706780 10 mon.a@0(leader).osd e47 remove_redundant_pg_temp
  95. 2012-11-21 11:07:03.478899 7f3db7706780 10 mon.a@0(leader).osd e47 update_logger
  96. 2012-11-21 11:07:03.478911 7f3db7706780 0 log [INF] : osdmap e47: 5 osds: 2 up, 2 in
  97. 2012-11-21 11:07:03.478916 7f3db7706780 1 -- 10.252.24.92:6789/0 --> mon.0 10.252.24.92:6789/0 -- log(2 entries) v1 -- ?+0 0x2997800
  98. 2012-11-21 11:07:03.478939 7f3db7706780 10 mon.a@0(leader).log v16222 create_pending v 16223
  99. 2012-11-21 11:07:03.478947 7f3db7706780 10 mon.a@0(leader).monmap v1 create_pending monmap epoch 2
  100. 2012-11-21 11:07:03.478956 7f3db7706780 0 log [INF] : monmap e1: 1 mons at {a=10.252.24.92:6789/0}
  101. 2012-11-21 11:07:03.478960 7f3db7706780 1 -- 10.252.24.92:6789/0 --> mon.0 10.252.24.92:6789/0 -- log(2 entries) v1 -- ?+0 0x2997600
  102. 2012-11-21 11:07:03.478968 7f3db7706780 10 mon.a@0(leader).auth v298 update_from_paxos()
  103. 2012-11-21 11:07:03.478971 7f3db7706780 10 mon.a@0(leader).auth v298 create_pending v 299
  104. 2012-11-21 11:07:03.478972 7f3db7706780 10 mon.a@0(leader).auth v298 AuthMonitor::on_active()
  105. 2012-11-21 11:07:03.478980 7f3db7706780 10 mon.a@0(leader) e1 resend_routed_requests
  106. 2012-11-21 11:07:03.478982 7f3db7706780 10 mon.a@0(leader) e1 register_cluster_logger
  107. 2012-11-21 11:07:03.479000 7f3db11be700 10 mon.a@0(leader) e1 do not have session, making new one
  108. 2012-11-21 11:07:03.479009 7f3db11be700 10 mon.a@0(leader) e1 ms_dispatch new session MonSession: osd.1 10.252.24.92:6805/15504 is open for osd.1 10.252.24.92:6805/15504
  109. 2012-11-21 11:07:03.479017 7f3db11be700 10 mon.a@0(leader) e1 setting timeout on session
  110. 2012-11-21 11:07:03.479023 7f3db11be700 10 mon.a@0(leader).auth v298 update_from_paxos()
  111. 2012-11-21 11:07:03.479026 7f3db11be700 10 mon.a@0(leader).auth v298 preprocess_query auth(proto 0 26 bytes epoch 1) v1 from osd.1 10.252.24.92:6805/15504
  112. 2012-11-21 11:07:03.479034 7f3db11be700 10 mon.a@0(leader).auth v298 prep_auth() blob_size=26
  113. 2012-11-21 11:07:03.479044 7f3db11be700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6805/15504 -- auth_reply(proto 1 0 Success) v1 -- ?+0 0x2997e00 con 0x29b4a00
  114. 2012-11-21 11:07:03.479052 7f3db11be700 0 -- 10.252.24.92:6789/0 submit_message auth_reply(proto 1 0 Success) v1 remote, 10.252.24.92:6805/15504, failed lossy con, dropping message 0x2997e00
  115. 2012-11-21 11:07:03.479070 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x2997000 con 0x29b4000
  116. 2012-11-21 11:07:03.479079 7f3db11be700 10 mon.a@0(leader) e1 do not have session, making new one
  117. 2012-11-21 11:07:03.479082 7f3db11be700 10 mon.a@0(leader) e1 ms_dispatch new session MonSession: mon.0 10.252.24.92:6789/0 is open for mon.0 10.252.24.92:6789/0
  118. 2012-11-21 11:07:03.479088 7f3db11be700 5 mon.a@0(leader) e1 setting monitor caps on this connection
  119. 2012-11-21 11:07:03.479098 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_query log(1 entries) v1 from mon.0 10.252.24.92:6789/0
  120. 2012-11-21 11:07:03.479102 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_log log(1 entries) v1 from mon.0
  121. 2012-11-21 11:07:03.479108 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_update log(1 entries) v1 from mon.0 10.252.24.92:6789/0
  122. 2012-11-21 11:07:03.479113 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_log log(1 entries) v1 from mon.0
  123. 2012-11-21 11:07:03.479119 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478553 mon.0 10.252.24.92:6789/0 1 : [INF] mon.a@0 won leader election with quorum 0
  124. 2012-11-21 11:07:03.479138 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x2996e00 con 0x29b4000
  125. 2012-11-21 11:07:03.479157 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_query log(1 entries) v1 from mon.0 10.252.24.92:6789/0
  126. 2012-11-21 11:07:03.479163 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_log log(1 entries) v1 from mon.0
  127. 2012-11-21 11:07:03.479168 7f3db11be700 10 mon.a@0(leader).log v16222 nothing new
  128. 2012-11-21 11:07:03.479172 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(2 entries) v1 ==== 0+0+0 (0 0 0) 0x2997a00 con 0x29b4000
  129. 2012-11-21 11:07:03.479184 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_query log(2 entries) v1 from mon.0 10.252.24.92:6789/0
  130. 2012-11-21 11:07:03.479189 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_log log(2 entries) v1 from mon.0
  131. 2012-11-21 11:07:03.479195 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_update log(2 entries) v1 from mon.0 10.252.24.92:6789/0
  132. 2012-11-21 11:07:03.479199 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_log log(2 entries) v1 from mon.0
  133. 2012-11-21 11:07:03.479202 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478553 mon.0 10.252.24.92:6789/0 1 : [INF] mon.a@0 won leader election with quorum 0
  134. 2012-11-21 11:07:03.479210 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478666 mon.0 10.252.24.92:6789/0 2 : [INF] pgmap v14625: 384 pgs: 348 active+clean, 36 active+remapped; 1000 MB data, 131 GB used, 971 GB / 1161 GB avail
  135. 2012-11-21 11:07:03.479222 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(2 entries) v1 ==== 0+0+0 (0 0 0) 0x2997800 con 0x29b4000
  136. 2012-11-21 11:07:03.479235 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_query log(2 entries) v1 from mon.0 10.252.24.92:6789/0
  137. 2012-11-21 11:07:03.479240 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_log log(2 entries) v1 from mon.0
  138. 2012-11-21 11:07:03.479246 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_update log(2 entries) v1 from mon.0 10.252.24.92:6789/0
  139. 2012-11-21 11:07:03.479251 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_log log(2 entries) v1 from mon.0
  140. 2012-11-21 11:07:03.479260 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478666 mon.0 10.252.24.92:6789/0 2 : [INF] pgmap v14625: 384 pgs: 348 active+clean, 36 active+remapped; 1000 MB data, 131 GB used, 971 GB / 1161 GB avail
  141. 2012-11-21 11:07:03.479270 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478723 mon.0 10.252.24.92:6789/0 3 : [INF] mdsmap e38: 1/1/1 up {0=a=up:active}
  142. 2012-11-21 11:07:03.479283 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(2 entries) v1 ==== 0+0+0 (0 0 0) 0x2997600 con 0x29b4000
  143. 2012-11-21 11:07:03.479297 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_query log(2 entries) v1 from mon.0 10.252.24.92:6789/0
  144. 2012-11-21 11:07:03.479300 7f3db11be700 10 mon.a@0(leader).log v16222 preprocess_log log(2 entries) v1 from mon.0
  145. 2012-11-21 11:07:03.479305 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_update log(2 entries) v1 from mon.0 10.252.24.92:6789/0
  146. 2012-11-21 11:07:03.479308 7f3db11be700 10 mon.a@0(leader).log v16222 prepare_log log(2 entries) v1 from mon.0
  147. 2012-11-21 11:07:03.479309 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478723 mon.0 10.252.24.92:6789/0 3 : [INF] mdsmap e38: 1/1/1 up {0=a=up:active}
  148. 2012-11-21 11:07:03.479315 7f3db11be700 10 mon.a@0(leader).log v16222 logging 2012-11-21 11:07:03.478913 mon.0 10.252.24.92:6789/0 4 : [INF] osdmap e47: 5 osds: 2 up, 2 in
  149. 2012-11-21 11:07:03.479323 7f3db11be700 1 -- 10.252.24.92:6789/0 <== osd.0 10.252.24.92:6801/15428 1 ==== auth(proto 0 26 bytes epoch 1) v1 ==== 56+0+0 (2594425907 0 0) 0x2996800 con 0x29b4780
  150. 2012-11-21 11:07:03.479328 7f3db11be700 10 mon.a@0(leader) e1 do not have session, making new one
  151. 2012-11-21 11:07:03.479329 7f3db11be700 10 mon.a@0(leader) e1 ms_dispatch new session MonSession: osd.0 10.252.24.92:6801/15428 is open for osd.0 10.252.24.92:6801/15428
  152. 2012-11-21 11:07:03.479333 7f3db11be700 10 mon.a@0(leader) e1 setting timeout on session
  153. 2012-11-21 11:07:03.479337 7f3db11be700 10 mon.a@0(leader).auth v298 update_from_paxos()
  154. 2012-11-21 11:07:03.479338 7f3db11be700 10 mon.a@0(leader).auth v298 preprocess_query auth(proto 0 26 bytes epoch 1) v1 from osd.0 10.252.24.92:6801/15428
  155. 2012-11-21 11:07:03.479341 7f3db11be700 10 mon.a@0(leader).auth v298 prep_auth() blob_size=26
  156. 2012-11-21 11:07:03.479345 7f3db11be700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6801/15428 -- auth_reply(proto 1 0 Success) v1 -- ?+0 0x2997e00 con 0x29b4780
  157. 2012-11-21 11:07:03.479350 7f3db11be700 0 -- 10.252.24.92:6789/0 submit_message auth_reply(proto 1 0 Success) v1 remote, 10.252.24.92:6801/15428, failed lossy con, dropping message 0x2997e00
  158. 2012-11-21 11:07:03.479360 7f3db11be700 1 -- 10.252.24.92:6789/0 <== client.? 10.252.24.92:0/15522 1 ==== auth(proto 0 30 bytes epoch 1) v1 ==== 60+0+0 (807942897 0 0) 0x2996a00 con 0x29b48c0
  159. 2012-11-21 11:07:03.479365 7f3db11be700 10 mon.a@0(leader) e1 do not have session, making new one
  160. 2012-11-21 11:07:03.479366 7f3db11be700 10 mon.a@0(leader) e1 ms_dispatch new session MonSession: client.? 10.252.24.92:0/15522 is open for client.? 10.252.24.92:0/15522
  161. 2012-11-21 11:07:03.479370 7f3db11be700 10 mon.a@0(leader) e1 setting timeout on session
  162. 2012-11-21 11:07:03.479373 7f3db11be700 10 mon.a@0(leader).auth v298 update_from_paxos()
  163. 2012-11-21 11:07:03.479374 7f3db11be700 10 mon.a@0(leader).auth v298 preprocess_query auth(proto 0 30 bytes epoch 1) v1 from client.? 10.252.24.92:0/15522
  164. 2012-11-21 11:07:03.479377 7f3db11be700 10 mon.a@0(leader).auth v298 prep_auth() blob_size=30
  165. 2012-11-21 11:07:03.479380 7f3db11be700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:0/15522 -- auth_reply(proto 1 0 Success) v1 -- ?+0 0x2997e00 con 0x29b48c0
  166. 2012-11-21 11:07:03.479384 7f3db11be700 0 -- 10.252.24.92:6789/0 submit_message auth_reply(proto 1 0 Success) v1 remote, 10.252.24.92:0/15522, failed lossy con, dropping message 0x2997e00
  167. 2012-11-21 11:07:03.479392 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mds.0 10.252.24.92:6800/15350 1 ==== auth(proto 0 26 bytes epoch 1) v1 ==== 56+0+0 (740518103 0 0) 0x2996c00 con 0x29b4dc0
  168. 2012-11-21 11:07:03.479397 7f3db11be700 10 mon.a@0(leader) e1 do not have session, making new one
  169. 2012-11-21 11:07:03.479398 7f3db11be700 10 mon.a@0(leader) e1 ms_dispatch new session MonSession: mds.0 10.252.24.92:6800/15350 is open for mds.0 10.252.24.92:6800/15350
  170. 2012-11-21 11:07:03.479401 7f3db11be700 10 mon.a@0(leader) e1 setting timeout on session
  171. 2012-11-21 11:07:03.479410 7f3db11be700 10 mon.a@0(leader).auth v298 update_from_paxos()
  172. 2012-11-21 11:07:03.479411 7f3db11be700 10 mon.a@0(leader).auth v298 preprocess_query auth(proto 0 26 bytes epoch 1) v1 from mds.0 10.252.24.92:6800/15350
  173. 2012-11-21 11:07:03.479418 7f3db11be700 10 mon.a@0(leader).auth v298 prep_auth() blob_size=26
  174. 2012-11-21 11:07:03.479421 7f3db11be700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6800/15350 -- auth_reply(proto 1 0 Success) v1 -- ?+0 0x2997e00 con 0x29b4dc0
  175. 2012-11-21 11:07:03.479426 7f3db11be700 0 -- 10.252.24.92:6789/0 submit_message auth_reply(proto 1 0 Success) v1 remote, 10.252.24.92:6800/15350, failed lossy con, dropping message 0x2997e00
  176. 2012-11-21 11:07:03.529269 7f3db09bd700 10 mon.a@0(leader).log v16222 encode_pending v 16223
  177. 2012-11-21 11:07:03.529304 7f3db09bd700 10 mon.a@0(leader) e1 get_global_paxos_version first call this election epoch, starting from 31231
  178. 2012-11-21 11:07:03.679028 7f3db09bd700 7 mon.a@0(leader).log v16223 update_from_paxos applying incremental log 16223 2012-11-21 11:07:03.478553 mon.0 10.252.24.92:6789/0 1 : [INF] mon.a@0 won leader election with quorum 0
  179. 2012-11-21 11:07:03.679078 7f3db09bd700 7 mon.a@0(leader).log v16223 update_from_paxos applying incremental log 16223 2012-11-21 11:07:03.478666 mon.0 10.252.24.92:6789/0 2 : [INF] pgmap v14625: 384 pgs: 348 active+clean, 36 active+remapped; 1000 MB data, 131 GB used, 971 GB / 1161 GB avail
  180. 2012-11-21 11:07:03.679110 7f3db09bd700 7 mon.a@0(leader).log v16223 update_from_paxos applying incremental log 16223 2012-11-21 11:07:03.478723 mon.0 10.252.24.92:6789/0 3 : [INF] mdsmap e38: 1/1/1 up {0=a=up:active}
  181. 2012-11-21 11:07:03.679127 7f3db09bd700 7 mon.a@0(leader).log v16223 update_from_paxos applying incremental log 16223 2012-11-21 11:07:03.478913 mon.0 10.252.24.92:6789/0 4 : [INF] osdmap e47: 5 osds: 2 up, 2 in
  182. 2012-11-21 11:07:03.813181 7f3db20c6700 1 -- 10.252.24.92:6789/0 >> :/0 pipe(0x298b200 sd=9 :6789 pgs=0 cs=0 l=0).accept sd=9
  183. 2012-11-21 11:07:03.813239 7f3db20c6700 10 mon.a@0(leader) e1 ms_verify_authorizer 10.252.24.92:0/15522 client protocol 0
  184. 2012-11-21 11:07:03.813435 7f3db11be700 1 -- 10.252.24.92:6789/0 <== client.? 10.252.24.92:0/15522 1 ==== auth(proto 0 30 bytes epoch 1) v1 ==== 60+0+0 (807942897 0 0) 0x2a92000 con 0x29b4c80
  185. 2012-11-21 11:07:03.821041 7f3db09bd700 10 mon.a@0(leader).log v16223 check_subs
  186. 2012-11-21 11:07:03.821064 7f3db09bd700 10 mon.a@0(leader).log v16223 create_pending v 16224
  187. 2012-11-21 11:07:03.821069 7f3db09bd700 7 mon.a@0(leader).log v16223 _updated_log for mon.0 10.252.24.92:6789/0
  188. 2012-11-21 11:07:03.821084 7f3db09bd700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6789/0 -- log(last 1) v1 -- ?+0 0x2a94000 con 0x29b4000
  189. 2012-11-21 11:07:03.821098 7f3db09bd700 7 mon.a@0(leader).log v16223 _updated_log for mon.0 10.252.24.92:6789/0
  190. 2012-11-21 11:07:03.821105 7f3db09bd700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6789/0 -- log(last 2) v1 -- ?+0 0x2a94340 con 0x29b4000
  191. 2012-11-21 11:07:03.821114 7f3db09bd700 7 mon.a@0(leader).log v16223 _updated_log for mon.0 10.252.24.92:6789/0
  192. 2012-11-21 11:07:03.821119 7f3db09bd700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6789/0 -- log(last 3) v1 -- ?+0 0x2a941a0 con 0x29b4000
  193. 2012-11-21 11:07:03.821127 7f3db09bd700 7 mon.a@0(leader).log v16223 _updated_log for mon.0 10.252.24.92:6789/0
  194. 2012-11-21 11:07:03.821133 7f3db09bd700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:6789/0 -- log(last 4) v1 -- ?+0 0x2a94820 con 0x29b4000
  195. 2012-11-21 11:07:03.821154 7f3db11be700 10 mon.a@0(leader) e1 do not have session, making new one
  196. 2012-11-21 11:07:03.821166 7f3db11be700 10 mon.a@0(leader) e1 ms_dispatch new session MonSession: client.? 10.252.24.92:0/15522 is open for client.? 10.252.24.92:0/15522
  197. 2012-11-21 11:07:03.821175 7f3db11be700 10 mon.a@0(leader) e1 setting timeout on session
  198. 2012-11-21 11:07:03.821185 7f3db11be700 10 mon.a@0(leader).auth v298 update_from_paxos()
  199. 2012-11-21 11:07:03.821188 7f3db11be700 10 mon.a@0(leader).auth v298 preprocess_query auth(proto 0 30 bytes epoch 1) v1 from client.? 10.252.24.92:0/15522
  200. 2012-11-21 11:07:03.821201 7f3db11be700 10 mon.a@0(leader).auth v298 prep_auth() blob_size=30
  201. 2012-11-21 11:07:03.821211 7f3db11be700 1 -- 10.252.24.92:6789/0 --> 10.252.24.92:0/15522 -- auth_reply(proto 1 0 Success) v1 -- ?+0 0x2997e00 con 0x29b4c80
  202. 2012-11-21 11:07:03.821244 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(last 1) v1 ==== 0+0+0 (0 0 0) 0x2a94000 con 0x29b4000
  203. 2012-11-21 11:07:03.821261 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(last 2) v1 ==== 0+0+0 (0 0 0) 0x2a94340 con 0x29b4000
  204. 2012-11-21 11:07:03.821271 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(last 3) v1 ==== 0+0+0 (0 0 0) 0x2a941a0 con 0x29b4000
  205. 2012-11-21 11:07:03.821279 7f3db11be700 1 -- 10.252.24.92:6789/0 <== mon.0 10.252.24.92:6789/0 0 ==== log(last 4) v1 ==== 0+0+0 (0 0 0) 0x2a94820 con 0x29b4000
  206. 2012-11-21 11:07:03.821475 7f3db11be700 1 -- 10.252.24.92:6789/0 <== client.? 10.252.24.92:0/15522 2 ==== mon_subscribe({monmap=2+}) v2 ==== 23+0+0 (2366214315 0 0) 0x2a1b500 con 0x29b4c80
  207. 2012-11-21 11:07:03.821498 7f3db11be700 10 mon.a@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+}) v2
  208. 2012-11-21 11:07:03.821508 7f3db11be700 10 mon.a@0(leader) e1 check_sub monmap next 2 have 1
  209. 2012-11-21 11:07:03.821511 7f3db11be700 1 -- 10.252.24.92:6789/0 --> client.? 10.252.24.92:0/15522 -- mon_subscribe_ack(300s) v1 -- ?+0 0x2a94820
  210. 2012-11-21 11:07:03.821529 7f3db11be700 1 -- 10.252.24.92:6789/0 <== client.? 10.252.24.92:0/15522 3 ==== mon_command(osd crush set 3 osd.3 1 pool=default rack=c row=c v 0) v1 ==== 115+0+0 (2151211406 0 0) 0x2a1b880 con 0x29b4c80
  211. 2012-11-21 11:07:03.821544 7f3db11be700 0 mon.a@0(leader) e1 handle_command mon_command(osd crush set 3 osd.3 1 pool=default rack=c row=c v 0) v1
  212. 2012-11-21 11:07:03.821557 7f3db11be700 10 mon.a@0(leader).osd e47 preprocess_query mon_command(osd crush set 3 osd.3 1 pool=default rack=c row=c v 0) v1 from client.? 10.252.24.92:0/15522
  213. 2012-11-21 11:07:03.821572 7f3db11be700 7 mon.a@0(leader).osd e47 prepare_update mon_command(osd crush set 3 osd.3 1 pool=default rack=c row=c v 0) v1 from client.? 10.252.24.92:0/15522
  214. 2012-11-21 11:07:03.821603 7f3db11be700 0 mon.a@0(leader).osd e47 adding/updating crush item id 3 name 'osd.3' weight 1 at location {pool=default,rack=c,row=c}
  215. Segmentation fault (core dumped)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement