diff mbox

[ceph-users] Ceph mds laggy and failed assert in function replay mds/journal.cc

Message ID CAAM7YAnQBwj=KSSfATeepbWyhOMAVwfoiC2AvkCEZrdPwkY2Ug@mail.gmail.com (mailing list archive)
State New, archived
Headers show

Commit Message

Yan, Zheng April 27, 2014, 4:04 a.m. UTC
On Sat, Apr 26, 2014 at 9:56 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
> Hi Greg,
>
> Actually our cluster is pretty empty, but we suspect we had a temporary
> network disconnection to one of our OSD, not sure if this caused the
> problem.
>
> Anyway we don't mind try the method you mentioned, how can we do that?
>

compile ceph-mds with the attached patch. add a line "mds
wipe_sessions = 1" to the ceph.conf,

Yan, Zheng

> Regards,
> Luke
>
>
> On Saturday, April 26, 2014, Gregory Farnum <greg@inktank.com> wrote:
>>
>> Hmm, it looks like your on-disk SessionMap is horrendously out of
>> date. Did your cluster get full at some point?
>>
>> In any case, we're working on tools to repair this now but they aren't
>> ready for use yet. Probably the only thing you could do is create an
>> empty sessionmap with a higher version than the ones the journal
>> refers to, but that might have other fallout effects...
>> -Greg
>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>
>>
>> On Fri, Apr 25, 2014 at 2:57 AM, Mohd Bazli Ab Karim
>> <bazli.abkarim@mimos.my> wrote:
>> > More logs. I ran ceph-mds  with debug-mds=20.
>> >
>> > -2> 2014-04-25 17:47:54.839672 7f0d6f3f0700 10 mds.0.journal
>> > EMetaBlob.replay inotable tablev 4316124 <= table 4317932
>> > -1> 2014-04-25 17:47:54.839674 7f0d6f3f0700 10 mds.0.journal
>> > EMetaBlob.replay sessionmap v8632368 -(1|2) == table 7239603 prealloc
>> > [1000041df86~1] used 1000041db9e
>> >   0> 2014-04-25 17:47:54.840733 7f0d6f3f0700 -1 mds/journal.cc: In
>> > function 'void EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)' thread
>> > 7f0d6f3f0700 time 2014-04-25 17:47:54.839688 mds/journal.cc: 1303: FAILED
>> > assert(session)
>> >
>> > Please look at the attachment for more details.
>> >
>> > Regards,
>> > Bazli
>> >
>> > From: Mohd Bazli Ab Karim
>> > Sent: Friday, April 25, 2014 12:26 PM
>> > To: 'ceph-devel@vger.kernel.org'; ceph-users@lists.ceph.com
>> > Subject: Ceph mds laggy and failed assert in function replay
>> > mds/journal.cc
>> >
>> > Dear Ceph-devel, ceph-users,
>> >
>> > I am currently facing issue with my ceph mds server. Ceph-mds daemon
>> > does not want to bring up back.
>> > Tried running that manually with ceph-mds -i mon01 -d but it shows that
>> > it stucks at failed assert(session) line 1303 in mds/journal.cc and aborted.
>> >
>> > Can someone shed some light in this issue.
>> > ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)
>> >
>> > Let me know if I need to send log with debug enabled.
>> >
>> > Regards,
>> > Bazli
>> >
>> > ________________________________
>> > DISCLAIMER:
>> >
>> >
>> > This e-mail (including any attachments) is for the addressee(s) only and
>> > may be confidential, especially as regards personal data. If you are not the
>> > intended recipient, please note that any dealing, review, distribution,
>> > printing, copying or use of this e-mail is strictly prohibited. If you have
>> > received this email in error, please notify the sender immediately and
>> > delete the original message (including any attachments).
>> >
>> >
>> > MIMOS Berhad is a research and development institution under the purview
>> > of the Malaysian Ministry of Science, Technology and Innovation. Opinions,
>> > conclusions and other information in this e-mail that do not relate to the
>> > official business of MIMOS Berhad and/or its subsidiaries shall be
>> > understood as neither given nor endorsed by MIMOS Berhad and/or its
>> > subsidiaries and neither MIMOS Berhad nor its subsidiaries accepts
>> > responsibility for the same. All liability arising from or in connection
>> > with computer viruses and/or corrupted e-mails is excluded to the fullest
>> > extent permitted by law.
>> >
>> >
>> > ------------------------------------------------------------------
>> > -
>> > -
>> > DISCLAIMER:
>> >
>> > This e-mail (including any attachments) is for the addressee(s)
>> > only and may contain confidential information. If you are not the
>> > intended recipient, please note that any dealing, review,
>> > distribution, printing, copying or use of this e-mail is strictly
>> > prohibited. If you have received this email in error, please notify
>> > the sender  immediately and delete the original message.
>> > MIMOS Berhad is a research and development institution under
>> > the purview of the Malaysian Ministry of Science, Technology and
>> > Innovation. Opinions, conclusions and other information in this e-
>> > mail that do not relate to the official business of MIMOS Berhad
>> > and/or its subsidiaries shall be understood as neither given nor
>> > endorsed by MIMOS Berhad and/or its subsidiaries and neither
>> > MIMOS Berhad nor its subsidiaries accepts responsibility for the
>> > same. All liability arising from or in connection with computer
>> > viruses and/or corrupted e-mails is excluded to the fullest extent
>> > permitted by law.
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>

Comments

Jingyuan Luke April 29, 2014, 3:24 a.m. UTC | #1
Hi,

We had applied the patch and recompile ceph as well as updated the
ceph.conf as per suggested, when we re-run ceph-mds we noticed the
following:


2014-04-29 10:45:22.260798 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51366457,12681393 no session for client.324186
2014-04-29 10:45:22.262419 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51366475,12681393 no session for client.324186
2014-04-29 10:45:22.267699 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51366665,12681393 no session for client.324186
2014-04-29 10:45:22.271664 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51366724,12681393 no session for client.324186
2014-04-29 10:45:22.281050 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51366945,12681393 no session for client.324186
2014-04-29 10:45:22.283196 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51366996,12681393 no session for client.324186
2014-04-29 10:45:22.287801 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51367043,12681393 no session for client.324186
2014-04-29 10:45:22.289967 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51367082,12681393 no session for client.324186
2014-04-29 10:45:22.291026 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51367110,12681393 no session for client.324186
2014-04-29 10:45:22.294459 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51367192,12681393 no session for client.324186
2014-04-29 10:45:22.297228 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51367257,12681393 no session for client.324186
2014-04-29 10:45:22.297477 7f90b971d700  0 log [WRN] :  replayed op
client.324186:51367264,12681393 no session for client.324186

tcmalloc: large alloc 1136660480 bytes == 0xb2019000 @  0x7f90c2564da7
0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed 0x7f90c231de9a
0x7f90c0cca3fd
tcmalloc: large alloc 2273316864 bytes == 0x15d73d000 @
0x7f90c2564da7 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed
0x7f90c231de9a 0x7f90c0cca3fd

ceph -s shows that MDS up:replay,

Also the messages above seemed to be repeating again after a while but
with a different session number. Is there a way for us to determine
that we are on the right track? Thanks.

Regards,
Luke

On Sun, Apr 27, 2014 at 12:04 PM, Yan, Zheng <ukernel@gmail.com> wrote:
> On Sat, Apr 26, 2014 at 9:56 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
>> Hi Greg,
>>
>> Actually our cluster is pretty empty, but we suspect we had a temporary
>> network disconnection to one of our OSD, not sure if this caused the
>> problem.
>>
>> Anyway we don't mind try the method you mentioned, how can we do that?
>>
>
> compile ceph-mds with the attached patch. add a line "mds
> wipe_sessions = 1" to the ceph.conf,
>
> Yan, Zheng
>
>> Regards,
>> Luke
>>
>>
>> On Saturday, April 26, 2014, Gregory Farnum <greg@inktank.com> wrote:
>>>
>>> Hmm, it looks like your on-disk SessionMap is horrendously out of
>>> date. Did your cluster get full at some point?
>>>
>>> In any case, we're working on tools to repair this now but they aren't
>>> ready for use yet. Probably the only thing you could do is create an
>>> empty sessionmap with a higher version than the ones the journal
>>> refers to, but that might have other fallout effects...
>>> -Greg
>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>
>>>
>>> On Fri, Apr 25, 2014 at 2:57 AM, Mohd Bazli Ab Karim
>>> <bazli.abkarim@mimos.my> wrote:
>>> > More logs. I ran ceph-mds  with debug-mds=20.
>>> >
>>> > -2> 2014-04-25 17:47:54.839672 7f0d6f3f0700 10 mds.0.journal
>>> > EMetaBlob.replay inotable tablev 4316124 <= table 4317932
>>> > -1> 2014-04-25 17:47:54.839674 7f0d6f3f0700 10 mds.0.journal
>>> > EMetaBlob.replay sessionmap v8632368 -(1|2) == table 7239603 prealloc
>>> > [1000041df86~1] used 1000041db9e
>>> >   0> 2014-04-25 17:47:54.840733 7f0d6f3f0700 -1 mds/journal.cc: In
>>> > function 'void EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)' thread
>>> > 7f0d6f3f0700 time 2014-04-25 17:47:54.839688 mds/journal.cc: 1303: FAILED
>>> > assert(session)
>>> >
>>> > Please look at the attachment for more details.
>>> >
>>> > Regards,
>>> > Bazli
>>> >
>>> > From: Mohd Bazli Ab Karim
>>> > Sent: Friday, April 25, 2014 12:26 PM
>>> > To: 'ceph-devel@vger.kernel.org'; ceph-users@lists.ceph.com
>>> > Subject: Ceph mds laggy and failed assert in function replay
>>> > mds/journal.cc
>>> >
>>> > Dear Ceph-devel, ceph-users,
>>> >
>>> > I am currently facing issue with my ceph mds server. Ceph-mds daemon
>>> > does not want to bring up back.
>>> > Tried running that manually with ceph-mds -i mon01 -d but it shows that
>>> > it stucks at failed assert(session) line 1303 in mds/journal.cc and aborted.
>>> >
>>> > Can someone shed some light in this issue.
>>> > ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)
>>> >
>>> > Let me know if I need to send log with debug enabled.
>>> >
>>> > Regards,
>>> > Bazli
>>> >
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Yan, Zheng April 29, 2014, 6:12 a.m. UTC | #2
On Tue, Apr 29, 2014 at 11:24 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
> Hi,
>
> We had applied the patch and recompile ceph as well as updated the
> ceph.conf as per suggested, when we re-run ceph-mds we noticed the
> following:
>
>
> 2014-04-29 10:45:22.260798 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51366457,12681393 no session for client.324186
> 2014-04-29 10:45:22.262419 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51366475,12681393 no session for client.324186
> 2014-04-29 10:45:22.267699 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51366665,12681393 no session for client.324186
> 2014-04-29 10:45:22.271664 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51366724,12681393 no session for client.324186
> 2014-04-29 10:45:22.281050 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51366945,12681393 no session for client.324186
> 2014-04-29 10:45:22.283196 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51366996,12681393 no session for client.324186
> 2014-04-29 10:45:22.287801 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51367043,12681393 no session for client.324186
> 2014-04-29 10:45:22.289967 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51367082,12681393 no session for client.324186
> 2014-04-29 10:45:22.291026 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51367110,12681393 no session for client.324186
> 2014-04-29 10:45:22.294459 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51367192,12681393 no session for client.324186
> 2014-04-29 10:45:22.297228 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51367257,12681393 no session for client.324186
> 2014-04-29 10:45:22.297477 7f90b971d700  0 log [WRN] :  replayed op
> client.324186:51367264,12681393 no session for client.324186
>
> tcmalloc: large alloc 1136660480 bytes == 0xb2019000 @  0x7f90c2564da7
> 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed 0x7f90c231de9a
> 0x7f90c0cca3fd
> tcmalloc: large alloc 2273316864 bytes == 0x15d73d000 @
> 0x7f90c2564da7 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed
> 0x7f90c231de9a 0x7f90c0cca3fd
>
> ceph -s shows that MDS up:replay,
>
> Also the messages above seemed to be repeating again after a while but
> with a different session number. Is there a way for us to determine
> that we are on the right track? Thanks.
>

It's on the right track as long as the MDS doesn't crash.

> Regards,
> Luke
>
> On Sun, Apr 27, 2014 at 12:04 PM, Yan, Zheng <ukernel@gmail.com> wrote:
>> On Sat, Apr 26, 2014 at 9:56 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
>>> Hi Greg,
>>>
>>> Actually our cluster is pretty empty, but we suspect we had a temporary
>>> network disconnection to one of our OSD, not sure if this caused the
>>> problem.
>>>
>>> Anyway we don't mind try the method you mentioned, how can we do that?
>>>
>>
>> compile ceph-mds with the attached patch. add a line "mds
>> wipe_sessions = 1" to the ceph.conf,
>>
>> Yan, Zheng
>>
>>> Regards,
>>> Luke
>>>
>>>
>>> On Saturday, April 26, 2014, Gregory Farnum <greg@inktank.com> wrote:
>>>>
>>>> Hmm, it looks like your on-disk SessionMap is horrendously out of
>>>> date. Did your cluster get full at some point?
>>>>
>>>> In any case, we're working on tools to repair this now but they aren't
>>>> ready for use yet. Probably the only thing you could do is create an
>>>> empty sessionmap with a higher version than the ones the journal
>>>> refers to, but that might have other fallout effects...
>>>> -Greg
>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>>
>>>>
>>>> On Fri, Apr 25, 2014 at 2:57 AM, Mohd Bazli Ab Karim
>>>> <bazli.abkarim@mimos.my> wrote:
>>>> > More logs. I ran ceph-mds  with debug-mds=20.
>>>> >
>>>> > -2> 2014-04-25 17:47:54.839672 7f0d6f3f0700 10 mds.0.journal
>>>> > EMetaBlob.replay inotable tablev 4316124 <= table 4317932
>>>> > -1> 2014-04-25 17:47:54.839674 7f0d6f3f0700 10 mds.0.journal
>>>> > EMetaBlob.replay sessionmap v8632368 -(1|2) == table 7239603 prealloc
>>>> > [1000041df86~1] used 1000041db9e
>>>> >   0> 2014-04-25 17:47:54.840733 7f0d6f3f0700 -1 mds/journal.cc: In
>>>> > function 'void EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)' thread
>>>> > 7f0d6f3f0700 time 2014-04-25 17:47:54.839688 mds/journal.cc: 1303: FAILED
>>>> > assert(session)
>>>> >
>>>> > Please look at the attachment for more details.
>>>> >
>>>> > Regards,
>>>> > Bazli
>>>> >
>>>> > From: Mohd Bazli Ab Karim
>>>> > Sent: Friday, April 25, 2014 12:26 PM
>>>> > To: 'ceph-devel@vger.kernel.org'; ceph-users@lists.ceph.com
>>>> > Subject: Ceph mds laggy and failed assert in function replay
>>>> > mds/journal.cc
>>>> >
>>>> > Dear Ceph-devel, ceph-users,
>>>> >
>>>> > I am currently facing issue with my ceph mds server. Ceph-mds daemon
>>>> > does not want to bring up back.
>>>> > Tried running that manually with ceph-mds -i mon01 -d but it shows that
>>>> > it stucks at failed assert(session) line 1303 in mds/journal.cc and aborted.
>>>> >
>>>> > Can someone shed some light in this issue.
>>>> > ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)
>>>> >
>>>> > Let me know if I need to send log with debug enabled.
>>>> >
>>>> > Regards,
>>>> > Bazli
>>>> >
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Jingyuan Luke April 29, 2014, 7:13 a.m. UTC | #3
Hi,

Assuming we got MDS working back on track, should we still leave the
mds_wipe_sessions in the ceph.conf or remove it and restart MDS.
Thanks.

Regards,
Luke


On Tue, Apr 29, 2014 at 2:12 PM, Yan, Zheng <ukernel@gmail.com> wrote:
> On Tue, Apr 29, 2014 at 11:24 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
>> Hi,
>>
>> We had applied the patch and recompile ceph as well as updated the
>> ceph.conf as per suggested, when we re-run ceph-mds we noticed the
>> following:
>>
>>
>> 2014-04-29 10:45:22.260798 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51366457,12681393 no session for client.324186
>> 2014-04-29 10:45:22.262419 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51366475,12681393 no session for client.324186
>> 2014-04-29 10:45:22.267699 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51366665,12681393 no session for client.324186
>> 2014-04-29 10:45:22.271664 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51366724,12681393 no session for client.324186
>> 2014-04-29 10:45:22.281050 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51366945,12681393 no session for client.324186
>> 2014-04-29 10:45:22.283196 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51366996,12681393 no session for client.324186
>> 2014-04-29 10:45:22.287801 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51367043,12681393 no session for client.324186
>> 2014-04-29 10:45:22.289967 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51367082,12681393 no session for client.324186
>> 2014-04-29 10:45:22.291026 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51367110,12681393 no session for client.324186
>> 2014-04-29 10:45:22.294459 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51367192,12681393 no session for client.324186
>> 2014-04-29 10:45:22.297228 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51367257,12681393 no session for client.324186
>> 2014-04-29 10:45:22.297477 7f90b971d700  0 log [WRN] :  replayed op
>> client.324186:51367264,12681393 no session for client.324186
>>
>> tcmalloc: large alloc 1136660480 bytes == 0xb2019000 @  0x7f90c2564da7
>> 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed 0x7f90c231de9a
>> 0x7f90c0cca3fd
>> tcmalloc: large alloc 2273316864 bytes == 0x15d73d000 @
>> 0x7f90c2564da7 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed
>> 0x7f90c231de9a 0x7f90c0cca3fd
>>
>> ceph -s shows that MDS up:replay,
>>
>> Also the messages above seemed to be repeating again after a while but
>> with a different session number. Is there a way for us to determine
>> that we are on the right track? Thanks.
>>
>
> It's on the right track as long as the MDS doesn't crash.
>
>> Regards,
>> Luke
>>
>> On Sun, Apr 27, 2014 at 12:04 PM, Yan, Zheng <ukernel@gmail.com> wrote:
>>> On Sat, Apr 26, 2014 at 9:56 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
>>>> Hi Greg,
>>>>
>>>> Actually our cluster is pretty empty, but we suspect we had a temporary
>>>> network disconnection to one of our OSD, not sure if this caused the
>>>> problem.
>>>>
>>>> Anyway we don't mind try the method you mentioned, how can we do that?
>>>>
>>>
>>> compile ceph-mds with the attached patch. add a line "mds
>>> wipe_sessions = 1" to the ceph.conf,
>>>
>>> Yan, Zheng
>>>
>>>> Regards,
>>>> Luke
>>>>
>>>>
>>>> On Saturday, April 26, 2014, Gregory Farnum <greg@inktank.com> wrote:
>>>>>
>>>>> Hmm, it looks like your on-disk SessionMap is horrendously out of
>>>>> date. Did your cluster get full at some point?
>>>>>
>>>>> In any case, we're working on tools to repair this now but they aren't
>>>>> ready for use yet. Probably the only thing you could do is create an
>>>>> empty sessionmap with a higher version than the ones the journal
>>>>> refers to, but that might have other fallout effects...
>>>>> -Greg
>>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>>>
>>>>>
>>>>> On Fri, Apr 25, 2014 at 2:57 AM, Mohd Bazli Ab Karim
>>>>> <bazli.abkarim@mimos.my> wrote:
>>>>> > More logs. I ran ceph-mds  with debug-mds=20.
>>>>> >
>>>>> > -2> 2014-04-25 17:47:54.839672 7f0d6f3f0700 10 mds.0.journal
>>>>> > EMetaBlob.replay inotable tablev 4316124 <= table 4317932
>>>>> > -1> 2014-04-25 17:47:54.839674 7f0d6f3f0700 10 mds.0.journal
>>>>> > EMetaBlob.replay sessionmap v8632368 -(1|2) == table 7239603 prealloc
>>>>> > [1000041df86~1] used 1000041db9e
>>>>> >   0> 2014-04-25 17:47:54.840733 7f0d6f3f0700 -1 mds/journal.cc: In
>>>>> > function 'void EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)' thread
>>>>> > 7f0d6f3f0700 time 2014-04-25 17:47:54.839688 mds/journal.cc: 1303: FAILED
>>>>> > assert(session)
>>>>> >
>>>>> > Please look at the attachment for more details.
>>>>> >
>>>>> > Regards,
>>>>> > Bazli
>>>>> >
>>>>> > From: Mohd Bazli Ab Karim
>>>>> > Sent: Friday, April 25, 2014 12:26 PM
>>>>> > To: 'ceph-devel@vger.kernel.org'; ceph-users@lists.ceph.com
>>>>> > Subject: Ceph mds laggy and failed assert in function replay
>>>>> > mds/journal.cc
>>>>> >
>>>>> > Dear Ceph-devel, ceph-users,
>>>>> >
>>>>> > I am currently facing issue with my ceph mds server. Ceph-mds daemon
>>>>> > does not want to bring up back.
>>>>> > Tried running that manually with ceph-mds -i mon01 -d but it shows that
>>>>> > it stucks at failed assert(session) line 1303 in mds/journal.cc and aborted.
>>>>> >
>>>>> > Can someone shed some light in this issue.
>>>>> > ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)
>>>>> >
>>>>> > Let me know if I need to send log with debug enabled.
>>>>> >
>>>>> > Regards,
>>>>> > Bazli
>>>>> >
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Yan, Zheng April 29, 2014, 7:36 a.m. UTC | #4
On Tue, Apr 29, 2014 at 3:13 PM, Jingyuan Luke <jyluke@gmail.com> wrote:
> Hi,
>
> Assuming we got MDS working back on track, should we still leave the
> mds_wipe_sessions in the ceph.conf or remove it and restart MDS.
> Thanks.

No.

It has been several hours. the MDS still does not finish replaying the journal?

Regards
Yan, Zheng

>
> Regards,
> Luke
>
>
> On Tue, Apr 29, 2014 at 2:12 PM, Yan, Zheng <ukernel@gmail.com> wrote:
>> On Tue, Apr 29, 2014 at 11:24 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
>>> Hi,
>>>
>>> We had applied the patch and recompile ceph as well as updated the
>>> ceph.conf as per suggested, when we re-run ceph-mds we noticed the
>>> following:
>>>
>>>
>>> 2014-04-29 10:45:22.260798 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51366457,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.262419 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51366475,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.267699 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51366665,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.271664 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51366724,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.281050 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51366945,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.283196 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51366996,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.287801 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51367043,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.289967 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51367082,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.291026 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51367110,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.294459 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51367192,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.297228 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51367257,12681393 no session for client.324186
>>> 2014-04-29 10:45:22.297477 7f90b971d700  0 log [WRN] :  replayed op
>>> client.324186:51367264,12681393 no session for client.324186
>>>
>>> tcmalloc: large alloc 1136660480 bytes == 0xb2019000 @  0x7f90c2564da7
>>> 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed 0x7f90c231de9a
>>> 0x7f90c0cca3fd
>>> tcmalloc: large alloc 2273316864 bytes == 0x15d73d000 @
>>> 0x7f90c2564da7 0x5bb9cb 0x5ac8eb 0x5b32f7 0x79ecd8 0x58cbed
>>> 0x7f90c231de9a 0x7f90c0cca3fd
>>>
>>> ceph -s shows that MDS up:replay,
>>>
>>> Also the messages above seemed to be repeating again after a while but
>>> with a different session number. Is there a way for us to determine
>>> that we are on the right track? Thanks.
>>>
>>
>> It's on the right track as long as the MDS doesn't crash.
>>
>>> Regards,
>>> Luke
>>>
>>> On Sun, Apr 27, 2014 at 12:04 PM, Yan, Zheng <ukernel@gmail.com> wrote:
>>>> On Sat, Apr 26, 2014 at 9:56 AM, Jingyuan Luke <jyluke@gmail.com> wrote:
>>>>> Hi Greg,
>>>>>
>>>>> Actually our cluster is pretty empty, but we suspect we had a temporary
>>>>> network disconnection to one of our OSD, not sure if this caused the
>>>>> problem.
>>>>>
>>>>> Anyway we don't mind try the method you mentioned, how can we do that?
>>>>>
>>>>
>>>> compile ceph-mds with the attached patch. add a line "mds
>>>> wipe_sessions = 1" to the ceph.conf,
>>>>
>>>> Yan, Zheng
>>>>
>>>>> Regards,
>>>>> Luke
>>>>>
>>>>>
>>>>> On Saturday, April 26, 2014, Gregory Farnum <greg@inktank.com> wrote:
>>>>>>
>>>>>> Hmm, it looks like your on-disk SessionMap is horrendously out of
>>>>>> date. Did your cluster get full at some point?
>>>>>>
>>>>>> In any case, we're working on tools to repair this now but they aren't
>>>>>> ready for use yet. Probably the only thing you could do is create an
>>>>>> empty sessionmap with a higher version than the ones the journal
>>>>>> refers to, but that might have other fallout effects...
>>>>>> -Greg
>>>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>>>>
>>>>>>
>>>>>> On Fri, Apr 25, 2014 at 2:57 AM, Mohd Bazli Ab Karim
>>>>>> <bazli.abkarim@mimos.my> wrote:
>>>>>> > More logs. I ran ceph-mds  with debug-mds=20.
>>>>>> >
>>>>>> > -2> 2014-04-25 17:47:54.839672 7f0d6f3f0700 10 mds.0.journal
>>>>>> > EMetaBlob.replay inotable tablev 4316124 <= table 4317932
>>>>>> > -1> 2014-04-25 17:47:54.839674 7f0d6f3f0700 10 mds.0.journal
>>>>>> > EMetaBlob.replay sessionmap v8632368 -(1|2) == table 7239603 prealloc
>>>>>> > [1000041df86~1] used 1000041db9e
>>>>>> >   0> 2014-04-25 17:47:54.840733 7f0d6f3f0700 -1 mds/journal.cc: In
>>>>>> > function 'void EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)' thread
>>>>>> > 7f0d6f3f0700 time 2014-04-25 17:47:54.839688 mds/journal.cc: 1303: FAILED
>>>>>> > assert(session)
>>>>>> >
>>>>>> > Please look at the attachment for more details.
>>>>>> >
>>>>>> > Regards,
>>>>>> > Bazli
>>>>>> >
>>>>>> > From: Mohd Bazli Ab Karim
>>>>>> > Sent: Friday, April 25, 2014 12:26 PM
>>>>>> > To: 'ceph-devel@vger.kernel.org'; ceph-users@lists.ceph.com
>>>>>> > Subject: Ceph mds laggy and failed assert in function replay
>>>>>> > mds/journal.cc
>>>>>> >
>>>>>> > Dear Ceph-devel, ceph-users,
>>>>>> >
>>>>>> > I am currently facing issue with my ceph mds server. Ceph-mds daemon
>>>>>> > does not want to bring up back.
>>>>>> > Tried running that manually with ceph-mds -i mon01 -d but it shows that
>>>>>> > it stucks at failed assert(session) line 1303 in mds/journal.cc and aborted.
>>>>>> >
>>>>>> > Can someone shed some light in this issue.
>>>>>> > ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)
>>>>>> >
>>>>>> > Let me know if I need to send log with debug enabled.
>>>>>> >
>>>>>> > Regards,
>>>>>> > Bazli
>>>>>> >
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Luke Jing Yuan April 29, 2014, 7:43 a.m. UTC | #5
Hi,

MDS did finish the replay and working after that but we are wondering should we leave the mds_wipe_sessions in ceph.conf or remove it.

Regards,
Luke

-----Original Message-----
From: ceph-users-bounces@lists.ceph.com [mailto:ceph-users-bounces@lists.ceph.com] On Behalf Of Yan, Zheng
Sent: Tuesday, 29 April, 2014 3:36 PM
To: Jingyuan Luke
Cc: ceph-devel@vger.kernel.org; ceph-users@lists.ceph.com
Subject: Re: [ceph-users] Ceph mds laggy and failed assert in function replay mds/journal.cc

On Tue, Apr 29, 2014 at 3:13 PM, Jingyuan Luke <jyluke@gmail.com> wrote:
> Hi,
>
> Assuming we got MDS working back on track, should we still leave the
> mds_wipe_sessions in the ceph.conf or remove it and restart MDS.
> Thanks.

No.

It has been several hours. the MDS still does not finish replaying the journal?

Regards
Yan, Zheng
Mohd Bazli Ab Karim April 30, 2014, 7:07 a.m. UTC | #6
Hi Zheng,

Sorry for the late reply. For sure, I will try this again after we completely verifying all content in the file system. Hopefully all will be good.
And, please confirm this, I will set debug_mds=10 for the ceph-mds, and do you want me to send the ceph-mon log too?

BTW, how to confirm that the mds has passed the beacon to mon or not?

Thank you so much Zheng!

Bazli

-----Original Message-----
From: Yan, Zheng [mailto:ukernel@gmail.com]
Sent: Tuesday, April 29, 2014 10:13 PM
To: Mohd Bazli Ab Karim
Cc: Luke Jing Yuan; Wong Ming Tat
Subject: Re: [ceph-users] Ceph mds laggy and failed assert in function replay mds/journal.cc

On Tue, Apr 29, 2014 at 5:30 PM, Mohd Bazli Ab Karim <bazli.abkarim@mimos.my> wrote:
> Hi Zheng,
>
> The another issue that Luke mentioned just now was like this.
> At first, we ran one mds (mon01) with the new compiled ceph-mds. It works fine with only one MDS running at that time. However, when we ran two more MDSes mon02 mon03 with the new compiled ceph-mds, it started acting weird.
> Mon01 which was became active at first, will have the error and started to respawning. Once respawning happened, mon03 will take over from mon01 as master mds, and replay happened again.
> Again, when mon03 became active, it will have the same error like below, and respawning again. So, it seems to me that replay will continue to happen from one mds to another when they got respawned.
>
> 2014-04-29 15:36:24.917798 7f5c36476700  1 mds.0.server
> reconnect_clients -- 1 sessions
> 2014-04-29 15:36:24.919620 7f5c2fb3e700  0 -- 10.4.118.23:6800/26401
> >> 10.1.64.181:0/1558263174 pipe(0x2924f5780 sd=41 :6800 s=0 pgs=0
> cs=0 l=0 c=0x37056e0).accept peer addr is really
> 10.1.64.181:0/1558263174 (socket is 10.1.64.181:57649/0)
> 2014-04-29 15:36:24.921661 7f5c36476700  0 log [DBG] : reconnect by
> client.884169 10.1.64.181:0/1558263174 after 0.003774
> 2014-04-29 15:36:24.921786 7f5c36476700  1 mds.0.12858 reconnect_done
> 2014-04-29 15:36:25.109391 7f5c36476700  1 mds.0.12858 handle_mds_map
> i am now mds.0.12858
> 2014-04-29 15:36:25.109413 7f5c36476700  1 mds.0.12858 handle_mds_map
> state change up:reconnect --> up:rejoin
> 2014-04-29 15:36:25.109417 7f5c36476700  1 mds.0.12858 rejoin_start
> 2014-04-29 15:36:26.918067 7f5c36476700  1 mds.0.12858
> rejoin_joint_start
> 2014-04-29 15:36:33.520985 7f5c36476700  1 mds.0.12858 rejoin_done
> 2014-04-29 15:36:36.252925 7f5c36476700  1 mds.0.12858 handle_mds_map
> i am now mds.0.12858
> 2014-04-29 15:36:36.252927 7f5c36476700  1 mds.0.12858 handle_mds_map
> state change up:rejoin --> up:active
> 2014-04-29 15:36:36.252932 7f5c36476700  1 mds.0.12858 recovery_done -- successful recovery!
> 2014-04-29 15:36:36.745833 7f5c36476700  1 mds.0.12858 active_start
> 2014-04-29 15:36:36.987854 7f5c36476700  1 mds.0.12858 cluster recovered.
> 2014-04-29 15:36:40.182604 7f5c36476700  0 mds.0.12858
> handle_mds_beacon no longer laggy
> 2014-04-29 15:36:57.947441 7f5c2fb3e700  0 -- 10.4.118.23:6800/26401
> >> 10.1.64.181:0/1558263174 pipe(0x2924f5780 sd=41 :6800 s=2 pgs=156
> cs=1 l=0 c=0x37056e0).fault with nothing to send, going to standby
> 2014-04-29 15:37:10.534593 7f5c36476700  1 mds.-1.-1 handle_mds_map i
> (10.4.118.23:6800/26401) dne in the mdsmap, respawning myself
> 2014-04-29 15:37:10.534604 7f5c36476700  1 mds.-1.-1 respawn
> 2014-04-29 15:37:10.534609 7f5c36476700  1 mds.-1.-1  e: '/usr/bin/ceph-mds'
> 2014-04-29 15:37:10.534612 7f5c36476700  1 mds.-1.-1  0: '/usr/bin/ceph-mds'
> 2014-04-29 15:37:10.534616 7f5c36476700  1 mds.-1.-1  1: '--cluster=ceph'
> 2014-04-29 15:37:10.534619 7f5c36476700  1 mds.-1.-1  2: '-i'
> 2014-04-29 15:37:10.534621 7f5c36476700  1 mds.-1.-1  3: 'mon03'
> 2014-04-29 15:37:10.534623 7f5c36476700  1 mds.-1.-1  4: '-f'
> 2014-04-29 15:37:10.534641 7f5c36476700  1 mds.-1.-1  cwd /
> 2014-04-29 15:37:12.155458 7f8907c8b780  0 ceph version  (), process
> ceph-mds, pid 26401
> 2014-04-29 15:37:12.249780 7f8902d10700  1 mds.-1.0 handle_mds_map
> standby
>
> p/s. we ran ceph-mon and ceph-mds on same servers, (mon01,mon02,mon03)
>
> I sent to you two log files, mon01 and mon03 where the scenario of mon03 have state->standby->replay->active->respawned. And also, mon01 which is now running as active as a single MDS at this moment.
>

After the MDS became ative, it did not send beacon to the monitor. It seems like the MDS was busy doing something else. If this issue still happen, set debug_mds=10 and send the log to me.

Regards
Yan, Zheng

> Regards,
> Bazli
> -----Original Message-----
> From: Luke Jing Yuan
> Sent: Tuesday, April 29, 2014 4:46 PM
> To: Yan, Zheng
> Cc: Mohd Bazli Ab Karim; Wong Ming Tat
> Subject: RE: [ceph-users] Ceph mds laggy and failed assert in function
> replay mds/journal.cc
>
> Hi Zheng,
>
> Thanks for the information. Actually we encounter another issue, in our original setup, we have 3 MDS running (say mon01, mon02 and mon03), when we do the replay/recovery we did it on mon01. After we completed, we restarted the mds again on mon02 and mon03 (without the mds_wipe_sessions and using the patched binary) but there was when we noticed something else, my colleague Mr. Bazli who started the original thread probably can explain a bit more on the observations made.
>
> Regards,
> Luke
>
> -----Original Message-----
> From: Yan, Zheng [mailto:ukernel@gmail.com]
> Sent: Tuesday, 29 April, 2014 4:26 PM
> To: Luke Jing Yuan
> Subject: Re: [ceph-users] Ceph mds laggy and failed assert in function
> replay mds/journal.cc
>
> On Tue, Apr 29, 2014 at 3:43 PM, Luke Jing Yuan <jyluke@mimos.my> wrote:
>> Hi,
>>
>> MDS did finish the replay and working after that but we are wondering should we leave the mds_wipe_sessions in ceph.conf or remove it.
>>
>
> should disable mds_wipe_sessions after mds starts working
>
>> Regards,
>> Luke
>>
>> -----Original Message-----
>> From: ceph-users-bounces@lists.ceph.com
>> [mailto:ceph-users-bounces@lists.ceph.com] On Behalf Of Yan, Zheng
>> Sent: Tuesday, 29 April, 2014 3:36 PM
>> To: Jingyuan Luke
>> Cc: ceph-devel@vger.kernel.org; ceph-users@lists.ceph.com
>> Subject: Re: [ceph-users] Ceph mds laggy and failed assert in
>> function replay mds/journal.cc
>>
>> On Tue, Apr 29, 2014 at 3:13 PM, Jingyuan Luke <jyluke@gmail.com> wrote:
>>> Hi,
>>>
>>> Assuming we got MDS working back on track, should we still leave the
>>> mds_wipe_sessions in the ceph.conf or remove it and restart MDS.
>>> Thanks.
>>
>> No.
>>
>> It has been several hours. the MDS still does not finish replaying the journal?
>>
>> Regards
>> Yan, Zheng
>>
>>
>>
>> ________________________________
>> DISCLAIMER:
>>
>>
>> This e-mail (including any attachments) is for the addressee(s) only and may be confidential, especially as regards personal data. If you are not the intended recipient, please note that any dealing, review, distribution, printing, copying or use of this e-mail is strictly prohibited. If you have received this email in error, please notify the sender immediately and delete the original message (including any attachments).
>>
>>
>> MIMOS Berhad is a research and development institution under the purview of the Malaysian Ministry of Science, Technology and Innovation. Opinions, conclusions and other information in this e-mail that do not relate to the official business of MIMOS Berhad and/or its subsidiaries shall be understood as neither given nor endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS Berhad nor its subsidiaries accepts responsibility for the same. All liability arising from or in connection with computer viruses and/or corrupted e-mails is excluded to the fullest extent permitted by law.
>>
>> ------------------------------------------------------------------
>> -
>> -
>> DISCLAIMER:
>>
>> This e-mail (including any attachments) is for the addressee(s) only
>> and may contain confidential information. If you are not the intended
>> recipient, please note that any dealing, review, distribution,
>> printing, copying or use of this e-mail is strictly prohibited. If
>> you have received this email in error, please notify the sender
>> immediately and delete the original message.
>> MIMOS Berhad is a research and development institution under the
>> purview of the Malaysian Ministry of Science, Technology and
>> Innovation. Opinions, conclusions and other information in this e-
>> mail that do not relate to the official business of MIMOS Berhad
>> and/or its subsidiaries shall be understood as neither given nor
>> endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS
>> Berhad nor its subsidiaries accepts responsibility for the same. All
>> liability arising from or in connection with computer viruses and/or
>> corrupted e-mails is excluded to the fullest extent permitted by law.
>>
>>
>
> ________________________________
> DISCLAIMER:
>
>
> This e-mail (including any attachments) is for the addressee(s) only and may be confidential, especially as regards personal data. If you are not the intended recipient, please note that any dealing, review, distribution, printing, copying or use of this e-mail is strictly prohibited. If you have received this email in error, please notify the sender immediately and delete the original message (including any attachments).
>
>
> MIMOS Berhad is a research and development institution under the purview of the Malaysian Ministry of Science, Technology and Innovation. Opinions, conclusions and other information in this e-mail that do not relate to the official business of MIMOS Berhad and/or its subsidiaries shall be understood as neither given nor endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS Berhad nor its subsidiaries accepts responsibility for the same. All liability arising from or in connection with computer viruses and/or corrupted e-mails is excluded to the fullest extent permitted by law.
>
>
> ------------------------------------------------------------------
> -
> -
> DISCLAIMER:
>
> This e-mail (including any attachments) is for the addressee(s) only
> and may contain confidential information. If you are not the intended
> recipient, please note that any dealing, review, distribution,
> printing, copying or use of this e-mail is strictly prohibited. If you
> have received this email in error, please notify the sender
> immediately and delete the original message.
> MIMOS Berhad is a research and development institution under the
> purview of the Malaysian Ministry of Science, Technology and
> Innovation. Opinions, conclusions and other information in this e-
> mail that do not relate to the official business of MIMOS Berhad
> and/or its subsidiaries shall be understood as neither given nor
> endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS
> Berhad nor its subsidiaries accepts responsibility for the same. All
> liability arising from or in connection with computer viruses and/or
> corrupted e-mails is excluded to the fullest extent permitted by law.
>
Yan, Zheng April 30, 2014, 3 p.m. UTC | #7
On Wed, Apr 30, 2014 at 3:07 PM, Mohd Bazli Ab Karim
<bazli.abkarim@mimos.my> wrote:
> Hi Zheng,
>
> Sorry for the late reply. For sure, I will try this again after we completely verifying all content in the file system. Hopefully all will be good.
> And, please confirm this, I will set debug_mds=10 for the ceph-mds, and do you want me to send the ceph-mon log too?

yes please.

>
> BTW, how to confirm that the mds has passed the beacon to mon or not?
>
read monitor's log

Regards
Yan, Zheng


> Thank you so much Zheng!
>
> Bazli
>
> -----Original Message-----
> From: Yan, Zheng [mailto:ukernel@gmail.com]
> Sent: Tuesday, April 29, 2014 10:13 PM
> To: Mohd Bazli Ab Karim
> Cc: Luke Jing Yuan; Wong Ming Tat
> Subject: Re: [ceph-users] Ceph mds laggy and failed assert in function replay mds/journal.cc
>
> On Tue, Apr 29, 2014 at 5:30 PM, Mohd Bazli Ab Karim <bazli.abkarim@mimos.my> wrote:
>> Hi Zheng,
>>
>> The another issue that Luke mentioned just now was like this.
>> At first, we ran one mds (mon01) with the new compiled ceph-mds. It works fine with only one MDS running at that time. However, when we ran two more MDSes mon02 mon03 with the new compiled ceph-mds, it started acting weird.
>> Mon01 which was became active at first, will have the error and started to respawning. Once respawning happened, mon03 will take over from mon01 as master mds, and replay happened again.
>> Again, when mon03 became active, it will have the same error like below, and respawning again. So, it seems to me that replay will continue to happen from one mds to another when they got respawned.
>>
>> 2014-04-29 15:36:24.917798 7f5c36476700  1 mds.0.server
>> reconnect_clients -- 1 sessions
>> 2014-04-29 15:36:24.919620 7f5c2fb3e700  0 -- 10.4.118.23:6800/26401
>> >> 10.1.64.181:0/1558263174 pipe(0x2924f5780 sd=41 :6800 s=0 pgs=0
>> cs=0 l=0 c=0x37056e0).accept peer addr is really
>> 10.1.64.181:0/1558263174 (socket is 10.1.64.181:57649/0)
>> 2014-04-29 15:36:24.921661 7f5c36476700  0 log [DBG] : reconnect by
>> client.884169 10.1.64.181:0/1558263174 after 0.003774
>> 2014-04-29 15:36:24.921786 7f5c36476700  1 mds.0.12858 reconnect_done
>> 2014-04-29 15:36:25.109391 7f5c36476700  1 mds.0.12858 handle_mds_map
>> i am now mds.0.12858
>> 2014-04-29 15:36:25.109413 7f5c36476700  1 mds.0.12858 handle_mds_map
>> state change up:reconnect --> up:rejoin
>> 2014-04-29 15:36:25.109417 7f5c36476700  1 mds.0.12858 rejoin_start
>> 2014-04-29 15:36:26.918067 7f5c36476700  1 mds.0.12858
>> rejoin_joint_start
>> 2014-04-29 15:36:33.520985 7f5c36476700  1 mds.0.12858 rejoin_done
>> 2014-04-29 15:36:36.252925 7f5c36476700  1 mds.0.12858 handle_mds_map
>> i am now mds.0.12858
>> 2014-04-29 15:36:36.252927 7f5c36476700  1 mds.0.12858 handle_mds_map
>> state change up:rejoin --> up:active
>> 2014-04-29 15:36:36.252932 7f5c36476700  1 mds.0.12858 recovery_done -- successful recovery!
>> 2014-04-29 15:36:36.745833 7f5c36476700  1 mds.0.12858 active_start
>> 2014-04-29 15:36:36.987854 7f5c36476700  1 mds.0.12858 cluster recovered.
>> 2014-04-29 15:36:40.182604 7f5c36476700  0 mds.0.12858
>> handle_mds_beacon no longer laggy
>> 2014-04-29 15:36:57.947441 7f5c2fb3e700  0 -- 10.4.118.23:6800/26401
>> >> 10.1.64.181:0/1558263174 pipe(0x2924f5780 sd=41 :6800 s=2 pgs=156
>> cs=1 l=0 c=0x37056e0).fault with nothing to send, going to standby
>> 2014-04-29 15:37:10.534593 7f5c36476700  1 mds.-1.-1 handle_mds_map i
>> (10.4.118.23:6800/26401) dne in the mdsmap, respawning myself
>> 2014-04-29 15:37:10.534604 7f5c36476700  1 mds.-1.-1 respawn
>> 2014-04-29 15:37:10.534609 7f5c36476700  1 mds.-1.-1  e: '/usr/bin/ceph-mds'
>> 2014-04-29 15:37:10.534612 7f5c36476700  1 mds.-1.-1  0: '/usr/bin/ceph-mds'
>> 2014-04-29 15:37:10.534616 7f5c36476700  1 mds.-1.-1  1: '--cluster=ceph'
>> 2014-04-29 15:37:10.534619 7f5c36476700  1 mds.-1.-1  2: '-i'
>> 2014-04-29 15:37:10.534621 7f5c36476700  1 mds.-1.-1  3: 'mon03'
>> 2014-04-29 15:37:10.534623 7f5c36476700  1 mds.-1.-1  4: '-f'
>> 2014-04-29 15:37:10.534641 7f5c36476700  1 mds.-1.-1  cwd /
>> 2014-04-29 15:37:12.155458 7f8907c8b780  0 ceph version  (), process
>> ceph-mds, pid 26401
>> 2014-04-29 15:37:12.249780 7f8902d10700  1 mds.-1.0 handle_mds_map
>> standby
>>
>> p/s. we ran ceph-mon and ceph-mds on same servers, (mon01,mon02,mon03)
>>
>> I sent to you two log files, mon01 and mon03 where the scenario of mon03 have state->standby->replay->active->respawned. And also, mon01 which is now running as active as a single MDS at this moment.
>>
>
> After the MDS became ative, it did not send beacon to the monitor. It seems like the MDS was busy doing something else. If this issue still happen, set debug_mds=10 and send the log to me.
>
> Regards
> Yan, Zheng
>
>> Regards,
>> Bazli
>> -----Original Message-----
>> From: Luke Jing Yuan
>> Sent: Tuesday, April 29, 2014 4:46 PM
>> To: Yan, Zheng
>> Cc: Mohd Bazli Ab Karim; Wong Ming Tat
>> Subject: RE: [ceph-users] Ceph mds laggy and failed assert in function
>> replay mds/journal.cc
>>
>> Hi Zheng,
>>
>> Thanks for the information. Actually we encounter another issue, in our original setup, we have 3 MDS running (say mon01, mon02 and mon03), when we do the replay/recovery we did it on mon01. After we completed, we restarted the mds again on mon02 and mon03 (without the mds_wipe_sessions and using the patched binary) but there was when we noticed something else, my colleague Mr. Bazli who started the original thread probably can explain a bit more on the observations made.
>>
>> Regards,
>> Luke
>>
>> -----Original Message-----
>> From: Yan, Zheng [mailto:ukernel@gmail.com]
>> Sent: Tuesday, 29 April, 2014 4:26 PM
>> To: Luke Jing Yuan
>> Subject: Re: [ceph-users] Ceph mds laggy and failed assert in function
>> replay mds/journal.cc
>>
>> On Tue, Apr 29, 2014 at 3:43 PM, Luke Jing Yuan <jyluke@mimos.my> wrote:
>>> Hi,
>>>
>>> MDS did finish the replay and working after that but we are wondering should we leave the mds_wipe_sessions in ceph.conf or remove it.
>>>
>>
>> should disable mds_wipe_sessions after mds starts working
>>
>>> Regards,
>>> Luke
>>>
>>> -----Original Message-----
>>> From: ceph-users-bounces@lists.ceph.com
>>> [mailto:ceph-users-bounces@lists.ceph.com] On Behalf Of Yan, Zheng
>>> Sent: Tuesday, 29 April, 2014 3:36 PM
>>> To: Jingyuan Luke
>>> Cc: ceph-devel@vger.kernel.org; ceph-users@lists.ceph.com
>>> Subject: Re: [ceph-users] Ceph mds laggy and failed assert in
>>> function replay mds/journal.cc
>>>
>>> On Tue, Apr 29, 2014 at 3:13 PM, Jingyuan Luke <jyluke@gmail.com> wrote:
>>>> Hi,
>>>>
>>>> Assuming we got MDS working back on track, should we still leave the
>>>> mds_wipe_sessions in the ceph.conf or remove it and restart MDS.
>>>> Thanks.
>>>
>>> No.
>>>
>>> It has been several hours. the MDS still does not finish replaying the journal?
>>>
>>> Regards
>>> Yan, Zheng
>>>
>>>
>>>
>>> ________________________________
>>> DISCLAIMER:
>>>
>>>
>>> This e-mail (including any attachments) is for the addressee(s) only and may be confidential, especially as regards personal data. If you are not the intended recipient, please note that any dealing, review, distribution, printing, copying or use of this e-mail is strictly prohibited. If you have received this email in error, please notify the sender immediately and delete the original message (including any attachments).
>>>
>>>
>>> MIMOS Berhad is a research and development institution under the purview of the Malaysian Ministry of Science, Technology and Innovation. Opinions, conclusions and other information in this e-mail that do not relate to the official business of MIMOS Berhad and/or its subsidiaries shall be understood as neither given nor endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS Berhad nor its subsidiaries accepts responsibility for the same. All liability arising from or in connection with computer viruses and/or corrupted e-mails is excluded to the fullest extent permitted by law.
>>>
>>> ------------------------------------------------------------------
>>> -
>>> -
>>> DISCLAIMER:
>>>
>>> This e-mail (including any attachments) is for the addressee(s) only
>>> and may contain confidential information. If you are not the intended
>>> recipient, please note that any dealing, review, distribution,
>>> printing, copying or use of this e-mail is strictly prohibited. If
>>> you have received this email in error, please notify the sender
>>> immediately and delete the original message.
>>> MIMOS Berhad is a research and development institution under the
>>> purview of the Malaysian Ministry of Science, Technology and
>>> Innovation. Opinions, conclusions and other information in this e-
>>> mail that do not relate to the official business of MIMOS Berhad
>>> and/or its subsidiaries shall be understood as neither given nor
>>> endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS
>>> Berhad nor its subsidiaries accepts responsibility for the same. All
>>> liability arising from or in connection with computer viruses and/or
>>> corrupted e-mails is excluded to the fullest extent permitted by law.
>>>
>>>
>>
>> ________________________________
>> DISCLAIMER:
>>
>>
>> This e-mail (including any attachments) is for the addressee(s) only and may be confidential, especially as regards personal data. If you are not the intended recipient, please note that any dealing, review, distribution, printing, copying or use of this e-mail is strictly prohibited. If you have received this email in error, please notify the sender immediately and delete the original message (including any attachments).
>>
>>
>> MIMOS Berhad is a research and development institution under the purview of the Malaysian Ministry of Science, Technology and Innovation. Opinions, conclusions and other information in this e-mail that do not relate to the official business of MIMOS Berhad and/or its subsidiaries shall be understood as neither given nor endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS Berhad nor its subsidiaries accepts responsibility for the same. All liability arising from or in connection with computer viruses and/or corrupted e-mails is excluded to the fullest extent permitted by law.
>>
>>
>> ------------------------------------------------------------------
>> -
>> -
>> DISCLAIMER:
>>
>> This e-mail (including any attachments) is for the addressee(s) only
>> and may contain confidential information. If you are not the intended
>> recipient, please note that any dealing, review, distribution,
>> printing, copying or use of this e-mail is strictly prohibited. If you
>> have received this email in error, please notify the sender
>> immediately and delete the original message.
>> MIMOS Berhad is a research and development institution under the
>> purview of the Malaysian Ministry of Science, Technology and
>> Innovation. Opinions, conclusions and other information in this e-
>> mail that do not relate to the official business of MIMOS Berhad
>> and/or its subsidiaries shall be understood as neither given nor
>> endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS
>> Berhad nor its subsidiaries accepts responsibility for the same. All
>> liability arising from or in connection with computer viruses and/or
>> corrupted e-mails is excluded to the fullest extent permitted by law.
>>
>
> ________________________________
> DISCLAIMER:
>
>
> This e-mail (including any attachments) is for the addressee(s) only and may be confidential, especially as regards personal data. If you are not the intended recipient, please note that any dealing, review, distribution, printing, copying or use of this e-mail is strictly prohibited. If you have received this email in error, please notify the sender immediately and delete the original message (including any attachments).
>
>
> MIMOS Berhad is a research and development institution under the purview of the Malaysian Ministry of Science, Technology and Innovation. Opinions, conclusions and other information in this e-mail that do not relate to the official business of MIMOS Berhad and/or its subsidiaries shall be understood as neither given nor endorsed by MIMOS Berhad and/or its subsidiaries and neither MIMOS Berhad nor its subsidiaries accepts responsibility for the same. All liability arising from or in connection with computer viruses and/or corrupted e-mails is excluded to the fullest extent permitted by law.
>
> ------------------------------------------------------------------
> -
> -
> DISCLAIMER:
>
> This e-mail (including any attachments) is for the addressee(s)
> only and may contain confidential information. If you are not the
> intended recipient, please note that any dealing, review,
> distribution, printing, copying or use of this e-mail is strictly
> prohibited. If you have received this email in error, please notify
> the sender  immediately and delete the original message.
> MIMOS Berhad is a research and development institution under
> the purview of the Malaysian Ministry of Science, Technology and
> Innovation. Opinions, conclusions and other information in this e-
> mail that do not relate to the official business of MIMOS Berhad
> and/or its subsidiaries shall be understood as neither given nor
> endorsed by MIMOS Berhad and/or its subsidiaries and neither
> MIMOS Berhad nor its subsidiaries accepts responsibility for the
> same. All liability arising from or in connection with computer
> viruses and/or corrupted e-mails is excluded to the fullest extent
> permitted by law.
>
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
diff mbox

Patch

From 989084f9703ad6c74adc31abedbbb0e308425683 Mon Sep 17 00:00:00 2001
From: "Yan, Zheng" <zheng.z.yan@intel.com>
Date: Sun, 27 Apr 2014 11:53:13 +0800
Subject: [PATCH] ceph: skip bad sessionmap during journal replay

Signed-off-by: Yan, Zheng <zheng.z.yan@intel.com>
---
 src/mds/journal.cc | 59 +++++++++++++++++++++++++++++++++---------------------
 1 file changed, 36 insertions(+), 23 deletions(-)

diff --git a/src/mds/journal.cc b/src/mds/journal.cc
index 41a79f9..7c19636 100644
--- a/src/mds/journal.cc
+++ b/src/mds/journal.cc
@@ -1293,36 +1293,49 @@  void EMetaBlob::replay(MDS *mds, LogSegment *logseg, MDSlaveUpdate *slaveup)
     if (mds->sessionmap.version >= sessionmapv) {
       dout(10) << "EMetaBlob.replay sessionmap v " << sessionmapv
 	       << " <= table " << mds->sessionmap.version << dendl;
-    } else {
-      dout(10) << "EMetaBlob.replay sessionmap v" << sessionmapv
+    } else if (mds->sessionmap.version + 2 >= sessionmapv) {
+      dout(10) << "EMetaBlob.replay sessionmap v " << sessionmapv
 	       << " -(1|2) == table " << mds->sessionmap.version
 	       << " prealloc " << preallocated_inos
 	       << " used " << used_preallocated_ino
 	       << dendl;
       Session *session = mds->sessionmap.get_session(client_name);
-      assert(session);
-      dout(20) << " (session prealloc " << session->info.prealloc_inos << ")" << dendl;
-      if (used_preallocated_ino) {
-	if (session->info.prealloc_inos.empty()) {
-	  // HRM: badness in the journal
-	  mds->clog.warn() << " replayed op " << client_reqs << " on session for " << client_name
-			   << " with empty prealloc_inos\n";
-	} else {
-	  inodeno_t next = session->next_ino();
-	  inodeno_t i = session->take_ino(used_preallocated_ino);
-	  if (next != i)
-	    mds->clog.warn() << " replayed op " << client_reqs << " used ino " << i
-			     << " but session next is " << next << "\n";
-	  assert(i == used_preallocated_ino);
-	  session->info.used_inos.clear();
+      if (session) {
+	dout(20) << " (session prealloc " << session->info.prealloc_inos << ")" << dendl;
+	if (used_preallocated_ino) {
+	  if (session->info.prealloc_inos.empty()) {
+	    // HRM: badness in the journal
+	    mds->clog.warn() << " replayed op " << client_reqs << " on session for "
+			     << client_name << " with empty prealloc_inos\n";
+	  } else {
+	    inodeno_t next = session->next_ino();
+	    inodeno_t i = session->take_ino(used_preallocated_ino);
+	    if (next != i)
+	      mds->clog.warn() << " replayed op " << client_reqs << " used ino " << i
+			       << " but session next is " << next << "\n";
+	    assert(i == used_preallocated_ino);
+	    session->info.used_inos.clear();
+	  }
+	  mds->sessionmap.projected = ++mds->sessionmap.version;
 	}
-	mds->sessionmap.projected = ++mds->sessionmap.version;
-      }
-      if (preallocated_inos.size()) {
-	session->info.prealloc_inos.insert(preallocated_inos);
-	mds->sessionmap.projected = ++mds->sessionmap.version;
+	if (preallocated_inos.size()) {
+	  session->info.prealloc_inos.insert(preallocated_inos);
+	  mds->sessionmap.projected = ++mds->sessionmap.version;
+	}
+	assert(sessionmapv == mds->sessionmap.version);
+      } else {
+	mds->clog.warn() << " replayed op " << client_reqs << " no session for "
+			 << client_name << "\n";
+	mds->sessionmap.version = sessionmapv;
+	mds->sessionmap.projected = sessionmapv;
       }
-      assert(sessionmapv == mds->sessionmap.version);
+    } else {
+      mds->clog.error() << "journal replay sessionmap v " << sessionmapv
+			<< " -(1|2) > table " << mds->sessionmap.version << "\n";
+      assert(g_conf->mds_wipe_sessions);
+      mds->sessionmap.wipe();
+      mds->sessionmap.version = sessionmapv;
+      mds->sessionmap.projected = sessionmapv;
     }
   }
 
-- 
1.9.0