Is any errors disply when execute "ceph-deploy osd prepare" ? Best wishes, Mika
2014-10-31 17:36 GMT+08:00 Subhadip Bagui <[email protected]>: > Hi, > > Can anyone please help on this > > Regards, > Subhadip > > > ------------------------------------------------------------------------------------------------------------------- > > On Fri, Oct 31, 2014 at 12:51 AM, Subhadip Bagui <[email protected]> > wrote: > >> Hi, >> >> I'm new in ceph and tying to install the cluster. I'm using single server >> for mon and osd. I've create one partition with device /dev/vdb1 containing >> 100 gb with ext4 fs and trying to add as an OSD in ceph monitor. But >> whenever I'm trying to activate the partition as osd block device we are >> getting issue. The partition can't be mount with ceph default osd >> mountpoint. Please let me know what I'm missing >> >> [root@ceph-admin my-cluster]# *ceph-deploy osd activate ceph-admin:vdb1* >> [ceph_deploy.conf][DEBUG ] found configuration file at: >> /root/.cephdeploy.conf >> [ceph_deploy.cli][INFO ] Invoked (1.5.18): /usr/bin/ceph-deploy osd >> activate ceph-admin:vdb1 >> [ceph_deploy.osd][DEBUG ] Activating cluster ceph disks >> ceph-admin:/dev/vdb1: >> [ceph-admin][DEBUG ] connected to host: ceph-admin >> [ceph-admin][DEBUG ] detect platform information from remote host >> [ceph-admin][DEBUG ] detect machine type >> [ceph_deploy.osd][INFO ] Distro info: CentOS 6.5 Final >> [ceph_deploy.osd][DEBUG ] activating host ceph-admin disk /dev/vdb1 >> [ceph_deploy.osd][DEBUG ] will use init type: sysvinit >> [ceph-admin][INFO ] Running command: ceph-disk -v activate --mark-init >> sysvinit --mount /dev/vdb1 >> [ceph-admin][WARNIN] No data was received after 300 seconds, >> disconnecting... >> [ceph-admin][INFO ] checking OSD status... >> [ceph-admin][INFO ] Running command: ceph --cluster=ceph osd stat >> --format=json >> [ceph-admin][WARNIN] No data was received after 300 seconds, >> disconnecting... >> [ceph-admin][INFO ] Running command: chkconfig ceph on >> >> ---- >> >> [root@ceph-admin my-cluster]#* ceph status* >> >> 2014-10-30 20:40:32.102741 7fcc7c591700 0 -- :/1003242 >> >> 10.203.238.165:6789/0 pipe(0x7fcc780204b0 sd=3 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7fcc78020740).fault >> >> 2014-10-30 20:40:35.103348 7fcc7c490700 0 -- :/1003242 >> >> 10.203.238.165:6789/0 pipe(0x7fcc6c000c00 sd=3 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7fcc6c000e90).fault >> >> 2014-10-30 20:40:38.103994 7fcc7c591700 0 -- :/1003242 >> >> 10.203.238.165:6789/0 pipe(0x7fcc6c003010 sd=3 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7fcc6c0032a0).fault >> >> 2014-10-30 20:40:41.104498 7fcc7c490700 0 -- :/1003242 >> >> 10.203.238.165:6789/0 pipe(0x7fcc6c0039d0 sd=3 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7fcc6c003c60).fault >> >> Regards, >> Subhadip >> >> ------------------------------------------------------------------------------------------------------------------- >> > > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
