Hi Sage,
I have dropped all unset .. and even restarted the osd
No dice .. OSDs are still stuck .
--snip--
ceph daemon osd.0 status │rtt
min/avg/max/mdev = 0.095/0.120/0.236/0.015
ms
{ "cluster_fsid":
"99babb8f-c880-4b32-a227-94aa483d4871",
│[root@ceph2 ~]# ceph daemon osd.1
status
"osd_fsid":
"1ad28bde-c23c-44ba-a3b7-0aaaafd3372e", │{
"cluster_fsid":
"99babb8f-c880-4b32-a227-94aa483d4871",
"whoami":
0, │
"osd_fsid":
"becc3252-6977-47d6-87af-7b1337e591d8",
"state":
"booting",
│ "whoami":
1,
"oldest_map":
1, │
"state":
"booting",
"newest_map":
24, │
"oldest_map":
1,
"num_pgs":
0} │
"newest_map":
21,
--snip--
--snip--
ceph osd
tree
# id weight type name up/down reweight
-1 2 root default
-3 1 host ceph1
0 1 osd.0 down 0
-2 1 host ceph2
1 1 osd.1 down 0
--snip--
--snip--
ceph -s
cluster 2929fa80-0841-4cb6-a133-90b2098fc802
health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean
monmap e2: 3 mons at {ceph0=
10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0},
election epoch 50, quorum 0,1,2 ceph0,ceph1,ceph2
osdmap e24: 2 osds: 0 up, 0 in
pgmap v25: 192 pgs, 3 pools, 0 bytes data, 0 objects
0 kB used, 0 kB / 0 kB avail
192 creating
--snip--
On Sat, Jul 26, 2014 at 5:57 PM, Sage Weil <[email protected]> wrote:
> On Sat, 26 Jul 2014, 10 minus wrote:
> > Hi,
> >
> > I just setup a test ceph installation on 3 node Centos 6.5 .
> > two of the nodes are used for hosting osds and the third acts as mon .
> >
> > Please note I'm using LVM so had to set up the osd using the manual
> install
> > guide.
> >
> > --snip--
> > ceph -s
> > cluster 2929fa80-0841-4cb6-a133-90b2098fc802
> > health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean;
> > noup,nodown,noout flag(s) set
> > monmap e2: 3 mons at{ceph0=
> 10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0
> > }, election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2
> > osdmap e21: 2 osds: 0 up, 0 in
> > flags noup,nodown,noout
> ^^^^
>
> Do 'ceph osd unset noup' and they should start up. You likely also want
> to clear nodown and noout as well.
>
> sage
>
>
> > pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects
> > 0 kB used, 0 kB / 0 kB avail
> > 192 creating
> > --snip--
> >
> > osd tree
> >
> > --snip--
> > ceph osd tree
> > # id weight type name up/down reweight
> > -1 2 root default
> > -3 1 host ceph1
> > 0 1 osd.0 down 0
> > -2 1 host ceph2
> > 1 1 osd.1 down 0
> > --snip--
> >
> > --snip--
> > ceph daemon osd.0 status
> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
> > "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0aaaafd3372e",
> > "whoami": 0,
> > "state": "booting",
> > "oldest_map": 1,
> > "newest_map": 21,
> > "num_pgs": 0}
> >
> > --snip--
> >
> > --snip--
> > ceph daemon osd.1 status
> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
> > "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8",
> > "whoami": 1,
> > "state": "booting",
> > "oldest_map": 1,
> > "newest_map": 21,
> > "num_pgs": 0}
> > --snip--
> >
> > # Cpus are idling
> >
> > # does anybody know what is wrong
> >
> > Thanks in advance
> >
> >
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com