Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-09 Thread ST Wong (ITSC)
Hi all,

I’m testing failover behavior of mon/mgr by stopping the active one.  Found 
message at the end of “ceph –s” output:

  progress:
Rebalancing after osd.71 marked out
  [..]

Is this normal?

Thanks and Rgds.
/st
From: ceph-users  On Behalf Of ST Wong (ITSC)
Sent: Monday, July 8, 2019 9:28 AM
To: Erik McCormick ; Paul Emmerich 
; Alfredo Deza 
Cc: ceph-users 
Subject: Re: [ceph-users] ceph-volume failed after replacing disk

Thanks for all your help.

I’m just curious if I can re-use the same ID after disk crush since it seems I 
can do that according to the manual.It’s totally okay to use other ID ☺
Finally recreated the OSD without specifying OSD ID – it takes ID 71 again.

Thanks again.
Best Rgds,
/st Wong


From: ceph-users 
mailto:ceph-users-boun...@lists.ceph.com>> 
On Behalf Of Erik McCormick
Sent: Friday, July 5, 2019 9:41 PM
To: Paul Emmerich mailto:paul.emmer...@croit.io>>
Cc: ceph-users mailto:ceph-users@lists.ceph.com>>
Subject: Re: [ceph-users] ceph-volume failed after replacing disk

If you create the OSD without specifying an ID it will grab the lowest 
available one. Unless you have other gaps somewhere, that ID would probably be 
the one you just removed.

-Erik

On Fri, Jul 5, 2019, 9:19 AM Paul Emmerich 
mailto:paul.emmer...@croit.io>> wrote:

On Fri, Jul 5, 2019 at 2:17 PM Alfredo Deza 
mailto:ad...@redhat.com>> wrote:
On Fri, Jul 5, 2019 at 6:23 AM ST Wong (ITSC) 
mailto:s...@itsc.cuhk.edu.hk>> wrote:
>
> Hi,
>
>
>
> I target to run just destroy and re-use the ID as stated in manual but seems 
> not working.
>
> Seems I’m unable to re-use the ID ?

The OSD replacement guide does not mention anything about crush and
auth commands. I believe you are now in a situation where the ID is no
longer able to be re-used, and ceph-volume
will not create one for you when specifying it in the CLI.

I don't know why there is so much attachment to these ID numbers, why
is it desirable to have that 71 number back again?

it avoids unnecessary rebalances

>
>
>
> Thanks.
>
> /stwong
>
>
>
>
>
> From: Paul Emmerich mailto:paul.emmer...@croit.io>>
> Sent: Friday, July 5, 2019 5:54 PM
> To: ST Wong (ITSC) mailto:s...@itsc.cuhk.edu.hk>>
> Cc: Eugen Block mailto:ebl...@nde.ag>>; 
> ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
>
>
>
>
> On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC) 
> mailto:s...@itsc.cuhk.edu.hk>> wrote:
>
> Hi,
>
> Yes, I run the commands before:
>
> # ceph osd crush remove osd.71
> device 'osd.71' does not appear in the crush map
> # ceph auth del osd.71
> entity osd.71 does not exist
>
>
>
> which is probably the reason why you couldn't recycle the OSD ID.
>
>
>
> Either run just destroy and re-use the ID or run purge and not re-use the ID.
>
> Manually deleting auth and crush entries is no longer needed since purge was 
> introduced.
>
>
>
>
>
> Paul
>
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io<http://www.croit.io>
> Tel: +49 89 1896585 90
>
>
>
>
> Thanks.
> /stwong
>
> -Original Message-
> From: ceph-users 
> mailto:ceph-users-boun...@lists.ceph.com>> 
> On Behalf Of Eugen Block
> Sent: Friday, July 5, 2019 4:54 PM
> To: ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
> Hi,
>
> did you also remove that OSD from crush and also from auth before recreating 
> it?
>
> ceph osd crush remove osd.71
> ceph auth del osd.71
>
> Regards,
> Eugen
>
>
> Zitat von "ST Wong (ITSC)" 
> mailto:s...@itsc.cuhk.edu.hk>>:
>
> > Hi all,
> >
> > We replaced a faulty disk out of N OSD and tried to follow steps
> > according to "Replacing and OSD" in
> > http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> > but got error:
> >
> > # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
> > --bluestore --data /dev/data/lv01 --osd-id
> > 71 --block.db /dev/db/lv01
> > Running command: /bin/ceph-authtool --gen-print-key Running command:
> > /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
> > /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> > -->  RuntimeError: The osd ID 71 is already in use or does not exist.
> >
> > ceph -s still shows  N OSDS.   I then remove with &q

Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-07 Thread ST Wong (ITSC)
Thanks for all your help.

I’m just curious if I can re-use the same ID after disk crush since it seems I 
can do that according to the manual.It’s totally okay to use other ID ☺
Finally recreated the OSD without specifying OSD ID – it takes ID 71 again.

Thanks again.
Best Rgds,
/st Wong


From: ceph-users  On Behalf Of Erik McCormick
Sent: Friday, July 5, 2019 9:41 PM
To: Paul Emmerich 
Cc: ceph-users 
Subject: Re: [ceph-users] ceph-volume failed after replacing disk

If you create the OSD without specifying an ID it will grab the lowest 
available one. Unless you have other gaps somewhere, that ID would probably be 
the one you just removed.

-Erik

On Fri, Jul 5, 2019, 9:19 AM Paul Emmerich 
mailto:paul.emmer...@croit.io>> wrote:

On Fri, Jul 5, 2019 at 2:17 PM Alfredo Deza 
mailto:ad...@redhat.com>> wrote:
On Fri, Jul 5, 2019 at 6:23 AM ST Wong (ITSC) 
mailto:s...@itsc.cuhk.edu.hk>> wrote:
>
> Hi,
>
>
>
> I target to run just destroy and re-use the ID as stated in manual but seems 
> not working.
>
> Seems I’m unable to re-use the ID ?

The OSD replacement guide does not mention anything about crush and
auth commands. I believe you are now in a situation where the ID is no
longer able to be re-used, and ceph-volume
will not create one for you when specifying it in the CLI.

I don't know why there is so much attachment to these ID numbers, why
is it desirable to have that 71 number back again?

it avoids unnecessary rebalances

>
>
>
> Thanks.
>
> /stwong
>
>
>
>
>
> From: Paul Emmerich mailto:paul.emmer...@croit.io>>
> Sent: Friday, July 5, 2019 5:54 PM
> To: ST Wong (ITSC) mailto:s...@itsc.cuhk.edu.hk>>
> Cc: Eugen Block mailto:ebl...@nde.ag>>; 
> ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
>
>
>
>
> On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC) 
> mailto:s...@itsc.cuhk.edu.hk>> wrote:
>
> Hi,
>
> Yes, I run the commands before:
>
> # ceph osd crush remove osd.71
> device 'osd.71' does not appear in the crush map
> # ceph auth del osd.71
> entity osd.71 does not exist
>
>
>
> which is probably the reason why you couldn't recycle the OSD ID.
>
>
>
> Either run just destroy and re-use the ID or run purge and not re-use the ID.
>
> Manually deleting auth and crush entries is no longer needed since purge was 
> introduced.
>
>
>
>
>
> Paul
>
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io<http://www.croit.io>
> Tel: +49 89 1896585 90
>
>
>
>
> Thanks.
> /stwong
>
> -Original Message-
> From: ceph-users 
> mailto:ceph-users-boun...@lists.ceph.com>> 
> On Behalf Of Eugen Block
> Sent: Friday, July 5, 2019 4:54 PM
> To: ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
> Hi,
>
> did you also remove that OSD from crush and also from auth before recreating 
> it?
>
> ceph osd crush remove osd.71
> ceph auth del osd.71
>
> Regards,
> Eugen
>
>
> Zitat von "ST Wong (ITSC)" 
> mailto:s...@itsc.cuhk.edu.hk>>:
>
> > Hi all,
> >
> > We replaced a faulty disk out of N OSD and tried to follow steps
> > according to "Replacing and OSD" in
> > http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> > but got error:
> >
> > # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
> > --bluestore --data /dev/data/lv01 --osd-id
> > 71 --block.db /dev/db/lv01
> > Running command: /bin/ceph-authtool --gen-print-key Running command:
> > /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
> > /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> > -->  RuntimeError: The osd ID 71 is already in use or does not exist.
> >
> > ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".
> >  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd
> > ls".
> >
> > However, repeating the ceph-volume command still gets same error.
> > We're running CEPH 14.2.1.   I must have some steps missed.Would
> > anyone please help? Thanks a lot.
> >
> > Rgds,
> > /stwong
>
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
> http://lists.ceph.com/listin

Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread Erik McCormick
If you create the OSD without specifying an ID it will grab the lowest
available one. Unless you have other gaps somewhere, that ID would probably
be the one you just removed.

-Erik

On Fri, Jul 5, 2019, 9:19 AM Paul Emmerich  wrote:

>
> On Fri, Jul 5, 2019 at 2:17 PM Alfredo Deza  wrote:
>
>> On Fri, Jul 5, 2019 at 6:23 AM ST Wong (ITSC) 
>> wrote:
>> >
>> > Hi,
>> >
>> >
>> >
>> > I target to run just destroy and re-use the ID as stated in manual but
>> seems not working.
>> >
>> > Seems I’m unable to re-use the ID ?
>>
>> The OSD replacement guide does not mention anything about crush and
>> auth commands. I believe you are now in a situation where the ID is no
>> longer able to be re-used, and ceph-volume
>> will not create one for you when specifying it in the CLI.
>>
>> I don't know why there is so much attachment to these ID numbers, why
>> is it desirable to have that 71 number back again?
>>
>
> it avoids unnecessary rebalances
>
>
>> >
>> >
>> >
>> > Thanks.
>> >
>> > /stwong
>> >
>> >
>> >
>> >
>> >
>> > From: Paul Emmerich 
>> > Sent: Friday, July 5, 2019 5:54 PM
>> > To: ST Wong (ITSC) 
>> > Cc: Eugen Block ; ceph-users@lists.ceph.com
>> > Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>> >
>> >
>> >
>> >
>> >
>> > On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC) 
>> wrote:
>> >
>> > Hi,
>> >
>> > Yes, I run the commands before:
>> >
>> > # ceph osd crush remove osd.71
>> > device 'osd.71' does not appear in the crush map
>> > # ceph auth del osd.71
>> > entity osd.71 does not exist
>> >
>> >
>> >
>> > which is probably the reason why you couldn't recycle the OSD ID.
>> >
>> >
>> >
>> > Either run just destroy and re-use the ID or run purge and not re-use
>> the ID.
>> >
>> > Manually deleting auth and crush entries is no longer needed since
>> purge was introduced.
>> >
>> >
>> >
>> >
>> >
>> > Paul
>> >
>> >
>> > --
>> > Paul Emmerich
>> >
>> > Looking for help with your Ceph cluster? Contact us at https://croit.io
>> >
>> > croit GmbH
>> > Freseniusstr. 31h
>> > 81247 München
>> > www.croit.io
>> > Tel: +49 89 1896585 90
>> >
>> >
>> >
>> >
>> > Thanks.
>> > /stwong
>> >
>> > -Original Message-
>> > From: ceph-users  On Behalf Of
>> Eugen Block
>> > Sent: Friday, July 5, 2019 4:54 PM
>> > To: ceph-users@lists.ceph.com
>> > Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>> >
>> > Hi,
>> >
>> > did you also remove that OSD from crush and also from auth before
>> recreating it?
>> >
>> > ceph osd crush remove osd.71
>> > ceph auth del osd.71
>> >
>> > Regards,
>> > Eugen
>> >
>> >
>> > Zitat von "ST Wong (ITSC)" :
>> >
>> > > Hi all,
>> > >
>> > > We replaced a faulty disk out of N OSD and tried to follow steps
>> > > according to "Replacing and OSD" in
>> > > http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
>> > > but got error:
>> > >
>> > > # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
>> > > --bluestore --data /dev/data/lv01 --osd-id
>> > > 71 --block.db /dev/db/lv01
>> > > Running command: /bin/ceph-authtool --gen-print-key Running command:
>> > > /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
>> > > /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
>> > > -->  RuntimeError: The osd ID 71 is already in use or does not exist.
>> > >
>> > > ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".
>> > >  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd
>> > > ls".
>> > >
>> > > However, repeating the ceph-volume command still gets same error.
>> > > We're running CEPH 14.2.1.   I must have some steps missed.Would
>> > > anyone please help? Thanks a lot.
>> > >
>> > > Rgds,
>> > > /stwong
>> >
>> >
>> >
>> > ___
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> > ___
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> > ___
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread Paul Emmerich
On Fri, Jul 5, 2019 at 2:17 PM Alfredo Deza  wrote:

> On Fri, Jul 5, 2019 at 6:23 AM ST Wong (ITSC)  wrote:
> >
> > Hi,
> >
> >
> >
> > I target to run just destroy and re-use the ID as stated in manual but
> seems not working.
> >
> > Seems I’m unable to re-use the ID ?
>
> The OSD replacement guide does not mention anything about crush and
> auth commands. I believe you are now in a situation where the ID is no
> longer able to be re-used, and ceph-volume
> will not create one for you when specifying it in the CLI.
>
> I don't know why there is so much attachment to these ID numbers, why
> is it desirable to have that 71 number back again?
>

it avoids unnecessary rebalances


> >
> >
> >
> > Thanks.
> >
> > /stwong
> >
> >
> >
> >
> >
> > From: Paul Emmerich 
> > Sent: Friday, July 5, 2019 5:54 PM
> > To: ST Wong (ITSC) 
> > Cc: Eugen Block ; ceph-users@lists.ceph.com
> > Subject: Re: [ceph-users] ceph-volume failed after replacing disk
> >
> >
> >
> >
> >
> > On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC) 
> wrote:
> >
> > Hi,
> >
> > Yes, I run the commands before:
> >
> > # ceph osd crush remove osd.71
> > device 'osd.71' does not appear in the crush map
> > # ceph auth del osd.71
> > entity osd.71 does not exist
> >
> >
> >
> > which is probably the reason why you couldn't recycle the OSD ID.
> >
> >
> >
> > Either run just destroy and re-use the ID or run purge and not re-use
> the ID.
> >
> > Manually deleting auth and crush entries is no longer needed since purge
> was introduced.
> >
> >
> >
> >
> >
> > Paul
> >
> >
> > --
> > Paul Emmerich
> >
> > Looking for help with your Ceph cluster? Contact us at https://croit.io
> >
> > croit GmbH
> > Freseniusstr. 31h
> > 81247 München
> > www.croit.io
> > Tel: +49 89 1896585 90
> >
> >
> >
> >
> > Thanks.
> > /stwong
> >
> > -Original Message-
> > From: ceph-users  On Behalf Of Eugen
> Block
> > Sent: Friday, July 5, 2019 4:54 PM
> > To: ceph-users@lists.ceph.com
> > Subject: Re: [ceph-users] ceph-volume failed after replacing disk
> >
> > Hi,
> >
> > did you also remove that OSD from crush and also from auth before
> recreating it?
> >
> > ceph osd crush remove osd.71
> > ceph auth del osd.71
> >
> > Regards,
> > Eugen
> >
> >
> > Zitat von "ST Wong (ITSC)" :
> >
> > > Hi all,
> > >
> > > We replaced a faulty disk out of N OSD and tried to follow steps
> > > according to "Replacing and OSD" in
> > > http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> > > but got error:
> > >
> > > # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
> > > --bluestore --data /dev/data/lv01 --osd-id
> > > 71 --block.db /dev/db/lv01
> > > Running command: /bin/ceph-authtool --gen-print-key Running command:
> > > /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
> > > /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> > > -->  RuntimeError: The osd ID 71 is already in use or does not exist.
> > >
> > > ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".
> > >  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd
> > > ls".
> > >
> > > However, repeating the ceph-volume command still gets same error.
> > > We're running CEPH 14.2.1.   I must have some steps missed.Would
> > > anyone please help? Thanks a lot.
> > >
> > > Rgds,
> > > /stwong
> >
> >
> >
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread Alfredo Deza
On Fri, Jul 5, 2019 at 6:23 AM ST Wong (ITSC)  wrote:
>
> Hi,
>
>
>
> I target to run just destroy and re-use the ID as stated in manual but seems 
> not working.
>
> Seems I’m unable to re-use the ID ?

The OSD replacement guide does not mention anything about crush and
auth commands. I believe you are now in a situation where the ID is no
longer able to be re-used, and ceph-volume
will not create one for you when specifying it in the CLI.

I don't know why there is so much attachment to these ID numbers, why
is it desirable to have that 71 number back again?
>
>
>
> Thanks.
>
> /stwong
>
>
>
>
>
> From: Paul Emmerich 
> Sent: Friday, July 5, 2019 5:54 PM
> To: ST Wong (ITSC) 
> Cc: Eugen Block ; ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
>
>
>
>
> On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC)  wrote:
>
> Hi,
>
> Yes, I run the commands before:
>
> # ceph osd crush remove osd.71
> device 'osd.71' does not appear in the crush map
> # ceph auth del osd.71
> entity osd.71 does not exist
>
>
>
> which is probably the reason why you couldn't recycle the OSD ID.
>
>
>
> Either run just destroy and re-use the ID or run purge and not re-use the ID.
>
> Manually deleting auth and crush entries is no longer needed since purge was 
> introduced.
>
>
>
>
>
> Paul
>
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io
> Tel: +49 89 1896585 90
>
>
>
>
> Thanks.
> /stwong
>
> -Original Message-
> From: ceph-users  On Behalf Of Eugen Block
> Sent: Friday, July 5, 2019 4:54 PM
> To: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
> Hi,
>
> did you also remove that OSD from crush and also from auth before recreating 
> it?
>
> ceph osd crush remove osd.71
> ceph auth del osd.71
>
> Regards,
> Eugen
>
>
> Zitat von "ST Wong (ITSC)" :
>
> > Hi all,
> >
> > We replaced a faulty disk out of N OSD and tried to follow steps
> > according to "Replacing and OSD" in
> > http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> > but got error:
> >
> > # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
> > --bluestore --data /dev/data/lv01 --osd-id
> > 71 --block.db /dev/db/lv01
> > Running command: /bin/ceph-authtool --gen-print-key Running command:
> > /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
> > /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> > -->  RuntimeError: The osd ID 71 is already in use or does not exist.
> >
> > ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".
> >  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd
> > ls".
> >
> > However, repeating the ceph-volume command still gets same error.
> > We're running CEPH 14.2.1.   I must have some steps missed.Would
> > anyone please help? Thanks a lot.
> >
> > Rgds,
> > /stwong
>
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread ST Wong (ITSC)
Hi,

I target to run just destroy and re-use the ID as stated in manual but seems 
not working.
Seems I’m unable to re-use the ID ?

Thanks.
/stwong


From: Paul Emmerich 
Sent: Friday, July 5, 2019 5:54 PM
To: ST Wong (ITSC) 
Cc: Eugen Block ; ceph-users@lists.ceph.com
Subject: Re: [ceph-users] ceph-volume failed after replacing disk


On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC) 
mailto:s...@itsc.cuhk.edu.hk>> wrote:
Hi,

Yes, I run the commands before:

# ceph osd crush remove osd.71
device 'osd.71' does not appear in the crush map
# ceph auth del osd.71
entity osd.71 does not exist

which is probably the reason why you couldn't recycle the OSD ID.

Either run just destroy and re-use the ID or run purge and not re-use the ID.
Manually deleting auth and crush entries is no longer needed since purge was 
introduced.


Paul

--
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io<http://www.croit.io>
Tel: +49 89 1896585 90


Thanks.
/stwong

-Original Message-
From: ceph-users 
mailto:ceph-users-boun...@lists.ceph.com>> 
On Behalf Of Eugen Block
Sent: Friday, July 5, 2019 4:54 PM
To: ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
Subject: Re: [ceph-users] ceph-volume failed after replacing disk

Hi,

did you also remove that OSD from crush and also from auth before recreating it?

ceph osd crush remove osd.71
ceph auth del osd.71

Regards,
Eugen


Zitat von "ST Wong (ITSC)" 
mailto:s...@itsc.cuhk.edu.hk>>:

> Hi all,
>
> We replaced a faulty disk out of N OSD and tried to follow steps
> according to "Replacing and OSD" in
> http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> but got error:
>
> # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
> --bluestore --data /dev/data/lv01 --osd-id
> 71 --block.db /dev/db/lv01
> Running command: /bin/ceph-authtool --gen-print-key Running command:
> /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
> /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> -->  RuntimeError: The osd ID 71 is already in use or does not exist.
>
> ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".
>  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd
> ls".
>
> However, repeating the ceph-volume command still gets same error.
> We're running CEPH 14.2.1.   I must have some steps missed.Would
> anyone please help? Thanks a lot.
>
> Rgds,
> /stwong



___
ceph-users mailing list
ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread Paul Emmerich
On Fri, Jul 5, 2019 at 11:25 AM ST Wong (ITSC)  wrote:

> Hi,
>
> Yes, I run the commands before:
>
> # ceph osd crush remove osd.71
> device 'osd.71' does not appear in the crush map
> # ceph auth del osd.71
> entity osd.71 does not exist
>

which is probably the reason why you couldn't recycle the OSD ID.

Either run just destroy and re-use the ID or run purge and not re-use the
ID.
Manually deleting auth and crush entries is no longer needed since purge
was introduced.


Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90


>
> Thanks.
> /stwong
>
> -Original Message-
> From: ceph-users  On Behalf Of Eugen
> Block
> Sent: Friday, July 5, 2019 4:54 PM
> To: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] ceph-volume failed after replacing disk
>
> Hi,
>
> did you also remove that OSD from crush and also from auth before
> recreating it?
>
> ceph osd crush remove osd.71
> ceph auth del osd.71
>
> Regards,
> Eugen
>
>
> Zitat von "ST Wong (ITSC)" :
>
> > Hi all,
> >
> > We replaced a faulty disk out of N OSD and tried to follow steps
> > according to "Replacing and OSD" in
> > http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> > but got error:
> >
> > # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create
> > --bluestore --data /dev/data/lv01 --osd-id
> > 71 --block.db /dev/db/lv01
> > Running command: /bin/ceph-authtool --gen-print-key Running command:
> > /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
> > /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> > -->  RuntimeError: The osd ID 71 is already in use or does not exist.
> >
> > ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".
> >  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd
> > ls".
> >
> > However, repeating the ceph-volume command still gets same error.
> > We're running CEPH 14.2.1.   I must have some steps missed.Would
> > anyone please help? Thanks a lot.
> >
> > Rgds,
> > /stwong
>
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread ST Wong (ITSC)
Hi,

Yes, I run the commands before:

# ceph osd crush remove osd.71
device 'osd.71' does not appear in the crush map
# ceph auth del osd.71
entity osd.71 does not exist

Thanks.
/stwong

-Original Message-
From: ceph-users  On Behalf Of Eugen Block
Sent: Friday, July 5, 2019 4:54 PM
To: ceph-users@lists.ceph.com
Subject: Re: [ceph-users] ceph-volume failed after replacing disk

Hi,

did you also remove that OSD from crush and also from auth before recreating it?

ceph osd crush remove osd.71
ceph auth del osd.71

Regards,
Eugen


Zitat von "ST Wong (ITSC)" :

> Hi all,
>
> We replaced a faulty disk out of N OSD and tried to follow steps 
> according to "Replacing and OSD" in 
> http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,
> but got error:
>
> # ceph osd destroy 71--yes-i-really-mean-it # ceph-volume lvm create 
> --bluestore --data /dev/data/lv01 --osd-id
> 71 --block.db /dev/db/lv01
> Running command: /bin/ceph-authtool --gen-print-key Running command: 
> /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring 
> /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> -->  RuntimeError: The osd ID 71 is already in use or does not exist.
>
> ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".   
>  Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd 
> ls".
>
> However, repeating the ceph-volume command still gets same error.
> We're running CEPH 14.2.1.   I must have some steps missed.Would  
> anyone please help? Thanks a lot.
>
> Rgds,
> /stwong



___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-volume failed after replacing disk

2019-07-05 Thread Eugen Block

Hi,

did you also remove that OSD from crush and also from auth before  
recreating it?


ceph osd crush remove osd.71
ceph auth del osd.71

Regards,
Eugen


Zitat von "ST Wong (ITSC)" :


Hi all,

We replaced a faulty disk out of N OSD and tried to follow steps  
according to "Replacing and OSD" in  
http://docs.ceph.com/docs/nautilus/rados/operations/add-or-rm-osds/,  
but got error:


# ceph osd destroy 71--yes-i-really-mean-it
# ceph-volume lvm create --bluestore --data /dev/data/lv01 --osd-id  
71 --block.db /dev/db/lv01

Running command: /bin/ceph-authtool --gen-print-key
Running command: /bin/ceph --cluster ceph --name  
client.bootstrap-osd --keyring  
/var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json

-->  RuntimeError: The osd ID 71 is already in use or does not exist.

ceph -s still shows  N OSDS.   I then remove with "ceph osd rm 71".   
 Now "ceph -s" shows N-1 OSDS and id 71 doesn't appear in "ceph osd  
ls".


However, repeating the ceph-volume command still gets same error.
We're running CEPH 14.2.1.   I must have some steps missed.Would  
anyone please help? Thanks a lot.


Rgds,
/stwong




___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com