Yes, that's what I did - as long as you don't affect the OSD data/partitions, 
they should come up again just fine once you reinstall ceph. I expect once the 
mandatory switch to ceph-volume finally happens this process might get a little 
more complicated but for jewel you're still using ceph-disk so it's just using 
udev rules to identify the OSDs and it will bring them up at start.

Rich

On 17/04/18 15:50, Shain Miley wrote:
> Rich,
> 
> Thank you for the information.  Are you suggesting that I do a clean install 
> of the OS partition...reinstall the Ceph packages, etc...and then Ceph should 
> be able to find the OSDs and all the original data.
> 
> I did do something similar one time when we had a MBR failure on a single 
> node and it would no longer boot up..however that was quite some time ago.
> 
> As far as I remember Ceph found all the OSD data just fine and everything did 
> startup in a good state post OS reinstall. 
> 
> Thanks again for your help on this issue.
> 
> Shain
> 
> 
> On 04/17/2018 06:00 AM, Richard Hesketh wrote:
>> On 16/04/18 18:32, Shain Miley wrote:
>>> Hello,
>>>
>>> We are currently running Ceph Jewel (10.2.10) on Ubuntu 14.04 in 
>>> production.  We have been running into a kernel panic bug off an on for a 
>>> while and I am starting to look into upgrading as a possible solution.  We 
>>> are currently running version 4.4.0-31-generic kernel on these servers and 
>>> have run across this same issue on multiple nodes over the course of the 
>>> last 12 months.
>>>
>>> My 3 questions are as follows:
>>>
>>> 1)Are there any known issues upgrading a Jewel cluster to 16.04?
>>>
>>> 2)At this point is Luminous stable enough to consider upgrading as well?
>>>
>>> 3)If I did decide to upgrade both Ubuntu and Ceph..in which order should 
>>> the upgrade occur?
>>>
>>> Thanks in advance for any insight that you are able to provide me on this 
>>> issue!
>>>
>>> Shain
>> When Luminous came out I did a multipart upgrade where I took my cluster 
>> from 14.04 to 16.04 (HWE kernel line), updated from jewel to luminous and 
>> migrated OSDs from filestore to bluestore (in that order). I had absolutely 
>> no issues throughout the process. The only thing I would suggest is that you 
>> reinstall your nodes to 16.04 rather than release-upgrading - previous 
>> experience with trying to release-upgrade on other hosts was sometimes 
>> painful, rebuilding was easier.
>>
>> Rich

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to