Re: [ceph-users] Migrating data into a newer ceph instance

2015-08-26 Thread Chang, Fangzhe (Fangzhe)
Thanks, Luis.

The motivation for using the newer version is to keep up-to-date with Ceph 
development, since we suspect the old versioned radosgw could not be restarted 
possibly due to library mismatch.
Do you know whether the self-healing feature of ceph is applicable between 
different versions or not?

Fangzhe

From: Luis Periquito [mailto:periqu...@gmail.com]
Sent: Wednesday, August 26, 2015 10:11 AM
To: Chang, Fangzhe (Fangzhe)
Cc: ceph-users@lists.ceph.com
Subject: Re: [ceph-users] Migrating data into a newer ceph instance

I Would say the easiest way would be to leverage all the self-healing of ceph: 
add the new nodes to the old cluster, allow or force all the data to migrate 
between nodes, and then remove the old ones out.

Well to be fair you could probably just install radosgw on another node and use 
it as your gateway without the need to even create a new OSD node...

Or was there a reason to create a new cluster? I can tell you that one of the 
clusters I have has been around since bobtail, and now it's hammer...

On Wed, Aug 26, 2015 at 2:50 PM, Chang, Fangzhe (Fangzhe) 
fangzhe.ch...@alcatel-lucent.commailto:fangzhe.ch...@alcatel-lucent.com 
wrote:
Hi,

We have been running Ceph/Radosgw version 0.80.7 (Giant) and stored quite some 
amount of data in it. We are only using ceph as an object store via radosgw. 
Last week cheph-radosgw daemon suddenly refused to start (with logs only 
showing “initialization timeout” error on Centos 7).  This triggers me to 
install a newer instance --- Ceph/Radosgw version 0.94.2 (Hammer). The new 
instance has a different set of key rings by default. The next step is to have 
all the data migrated. Does anyone know how to get the existing data out of the 
old ceph  cluster (Giant) and into the new instance (Hammer)? Please note that 
in the old three-node cluster ceph osd is still running but radosgw is not. Any 
suggestion will be greatly appreciated.
Thanks.

Regards,

Fangzhe Chang




___
ceph-users mailing list
ceph-users@lists.ceph.commailto:ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Migrating data into a newer ceph instance

2015-08-26 Thread Chang, Fangzhe (Fangzhe)
Hi,

We have been running Ceph/Radosgw version 0.80.7 (Giant) and stored quite some 
amount of data in it. We are only using ceph as an object store via radosgw. 
Last week cheph-radosgw daemon suddenly refused to start (with logs only 
showing initialization timeout error on Centos 7).  This triggers me to 
install a newer instance --- Ceph/Radosgw version 0.94.2 (Hammer). The new 
instance has a different set of key rings by default. The next step is to have 
all the data migrated. Does anyone know how to get the existing data out of the 
old ceph  cluster (Giant) and into the new instance (Hammer)? Please note that 
in the old three-node cluster ceph osd is still running but radosgw is not. Any 
suggestion will be greatly appreciated.
Thanks.

Regards,

Fangzhe Chang



___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Migrating data into a newer ceph instance

2015-08-26 Thread Luis Periquito
I Would say the easiest way would be to leverage all the self-healing of
ceph: add the new nodes to the old cluster, allow or force all the data to
migrate between nodes, and then remove the old ones out.

Well to be fair you could probably just install radosgw on another node and
use it as your gateway without the need to even create a new OSD node...

Or was there a reason to create a new cluster? I can tell you that one of
the clusters I have has been around since bobtail, and now it's hammer...

On Wed, Aug 26, 2015 at 2:50 PM, Chang, Fangzhe (Fangzhe) 
fangzhe.ch...@alcatel-lucent.com wrote:

 Hi,



 We have been running Ceph/Radosgw version 0.80.7 (Giant) and stored quite
 some amount of data in it. We are only using ceph as an object store via
 radosgw. Last week cheph-radosgw daemon suddenly refused to start (with
 logs only showing “initialization timeout” error on Centos 7).  This
 triggers me to install a newer instance --- Ceph/Radosgw version 0.94.2
 (Hammer). The new instance has a different set of key rings by default. The
 next step is to have all the data migrated. Does anyone know how to get the
 existing data out of the old ceph  cluster (Giant) and into the new
 instance (Hammer)? Please note that in the old three-node cluster ceph osd
 is still running but radosgw is not. Any suggestion will be greatly
 appreciated.

 Thanks.



 Regards,



 Fangzhe Chang







 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Migrating data into a newer ceph instance

2015-08-26 Thread Luis Periquito
I tend to not do too much each time: either upgrade or data migrate. The
actual upgrade process is seamless... So you can just as easily upgrade the
current cluster to hammer, and add/remove nodes on the fly. All of this
is quite seamless and straightforward (other than the data migration
itself).

On Wed, Aug 26, 2015 at 3:17 PM, Chang, Fangzhe (Fangzhe) 
fangzhe.ch...@alcatel-lucent.com wrote:

 Thanks, Luis.



 The motivation for using the newer version is to keep up-to-date with Ceph
 development, since we suspect the old versioned radosgw could not be
 restarted possibly due to library mismatch.

 Do you know whether the self-healing feature of ceph is applicable between
 different versions or not?



 Fangzhe



 *From:* Luis Periquito [mailto:periqu...@gmail.com]
 *Sent:* Wednesday, August 26, 2015 10:11 AM
 *To:* Chang, Fangzhe (Fangzhe)
 *Cc:* ceph-users@lists.ceph.com
 *Subject:* Re: [ceph-users] Migrating data into a newer ceph instance



 I Would say the easiest way would be to leverage all the self-healing of
 ceph: add the new nodes to the old cluster, allow or force all the data to
 migrate between nodes, and then remove the old ones out.



 Well to be fair you could probably just install radosgw on another node
 and use it as your gateway without the need to even create a new OSD node...



 Or was there a reason to create a new cluster? I can tell you that one of
 the clusters I have has been around since bobtail, and now it's hammer...



 On Wed, Aug 26, 2015 at 2:50 PM, Chang, Fangzhe (Fangzhe) 
 fangzhe.ch...@alcatel-lucent.com wrote:

 Hi,



 We have been running Ceph/Radosgw version 0.80.7 (Giant) and stored quite
 some amount of data in it. We are only using ceph as an object store via
 radosgw. Last week cheph-radosgw daemon suddenly refused to start (with
 logs only showing “initialization timeout” error on Centos 7).  This
 triggers me to install a newer instance --- Ceph/Radosgw version 0.94.2
 (Hammer). The new instance has a different set of key rings by default. The
 next step is to have all the data migrated. Does anyone know how to get the
 existing data out of the old ceph  cluster (Giant) and into the new
 instance (Hammer)? Please note that in the old three-node cluster ceph osd
 is still running but radosgw is not. Any suggestion will be greatly
 appreciated.

 Thanks.



 Regards,



 Fangzhe Chang








 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com