[root@chefwks01 ~]# ceph --cluster us-west01 osd crush dump
{ "devices": [
{ "id": 0,
"name": "osd.0"},
{ "id": 1,
"name": "osd.1"},
{ "id": 2,
"name": "osd.2"},
{ "id": 3,
"name": "osd.3"},
{ "id": 4,
"name": "osd.4"}],
"types": [
{ "type_id": 0,
"name": "osd"},
{ "type_id": 1,
"name": "host"},
{ "type_id": 2,
"name": "chassis"},
{ "type_id": 3,
"name": "rack"},
{ "type_id": 4,
"name": "row"},
{ "type_id": 5,
"name": "pdu"},
{ "type_id": 6,
"name": "pod"},
{ "type_id": 7,
"name": "room"},
{ "type_id": 8,
"name": "datacenter"},
{ "type_id": 9,
"name": "region"},
{ "type_id": 10,
"name": "root"}],
"buckets": [
{ "id": -1,
"name": "default",
"type_id": 10,
"type_name": "root",
"weight": 147455,
"alg": "straw",
"hash": "rjenkins1",
"items": [
{ "id": -2,
"weight": 29491,
"pos": 0},
{ "id": -3,
"weight": 29491,
"pos": 1},
{ "id": -4,
"weight": 29491,
"pos": 2},
{ "id": -5,
"weight": 29491,
"pos": 3},
{ "id": -6,
"weight": 29491,
"pos": 4}]},
{ "id": -2,
"name": "ceph-node20",
"type_id": 1,
"type_name": "host",
"weight": 29491,
"alg": "straw",
"hash": "rjenkins1",
"items": [
{ "id": 0,
"weight": 29491,
"pos": 0}]},
{ "id": -3,
"name": "ceph-node22",
"type_id": 1,
"type_name": "host",
"weight": 29491,
"alg": "straw",
"hash": "rjenkins1",
"items": [
{ "id": 2,
"weight": 29491,
"pos": 0}]},
{ "id": -4,
"name": "ceph-node24",
"type_id": 1,
"type_name": "host",
"weight": 29491,
"alg": "straw",
"hash": "rjenkins1",
"items": [
{ "id": 4,
"weight": 29491,
"pos": 0}]},
{ "id": -5,
"name": "ceph-node21",
"type_id": 1,
"type_name": "host",
"weight": 29491,
"alg": "straw",
"hash": "rjenkins1",
"items": [
{ "id": 1,
"weight": 29491,
"pos": 0}]},
{ "id": -6,
"name": "ceph-node23",
"type_id": 1,
"type_name": "host",
"weight": 29491,
"alg": "straw",
"hash": "rjenkins1",
"items": [
{ "id": 3,
"weight": 29491,
"pos": 0}]}],
"rules": [
{ "rule_id": 0,
"rule_name": "replicated_ruleset",
"ruleset": 0,
"type": 1,
"min_size": 1,
"max_size": 10,
"steps": [
{ "op": "take",
"item": -1,
"item_name": "default"},
{ "op": "chooseleaf_firstn",
"num": 0,
"type": "host"},
{ "op": "emit"}]},
{ "rule_id": 1,
"rule_name": "erasure-code",
"ruleset": 1,
"type": 3,
"min_size": 3,
"max_size": 20,
"steps": [
{ "op": "set_chooseleaf_tries",
"num": 5},
{ "op": "take",
"item": -1,
"item_name": "default"},
{ "op": "chooseleaf_indep",
"num": 0,
"type": "host"},
{ "op": "emit"}]},
{ "rule_id": 2,
"rule_name": "ecpool",
"ruleset": 2,
"type": 3,
"min_size": 3,
"max_size": 20,
"steps": [
{ "op": "set_chooseleaf_tries",
"num": 5},
{ "op": "take",
"item": -1,
"item_name": "default"},
{ "op": "choose_indep",
"num": 0,
"type": "osd"},
{ "op": "emit"}]}],
"tunables": { "choose_local_tries": 0,
"choose_local_fallback_tries": 0,
"choose_total_tries": 50,
"chooseleaf_descend_once": 1,
"profile": "bobtail",
"optimal_tunables": 0,
"legacy_tunables": 0,
"require_feature_tunables": 1,
"require_feature_tunables2": 1}}
On Thu, Jul 10, 2014 at 8:16 PM, Sage Weil <[email protected]> wrote:
> That is CEPH_FEATURE_CRUSH_V2. Can you attach teh output of
>
> ceph osd crush dump
>
> Thanks!
> sage
>
>
> On Thu, 10 Jul 2014, Joshua McClintock wrote:
>
> > Yes, I change some of the mount options on my osds (xfs mount options),
> but
> > I think this may be the answer from dmesg, sorta looks like a version
> > mismatch:
> >
> > libceph: loaded (mon/osd proto 15/24)
> >
> > ceph: loaded (mds proto 32)
> >
> > libceph: mon0 192.168.0.14:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon0 192.168.0.14:6789 socket error on read
> >
> > libceph: mon2 192.168.0.16:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon2 192.168.0.16:6789 socket error on read
> >
> > libceph: mon1 192.168.0.15:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon1 192.168.0.15:6789 socket error on read
> >
> > libceph: mon0 192.168.0.14:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon0 192.168.0.14:6789 socket error on read
> >
> > libceph: mon2 192.168.0.16:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon2 192.168.0.16:6789 socket error on read
> >
> > libceph: mon1 192.168.0.15:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon1 192.168.0.15:6789 socket error on read
> >
> >
> > I maybe I didn't update as well as I thought it did. I did hit every
> mon,
> > but I remember I couldn't upgrade to the new 'ceph' package because it
> > conflicted with 'python-ceph', so I uninstalled it (python-ceph), and
> then
> > upgraded to .80.1-2. Maybe there's a subcomponent I missed?
> >
> >
> > Here's rpm -qa from the client:
> >
> >
> > [root@chefwks01 ~]# rpm -qa|grep ceph
> >
> > ceph-deploy-1.5.2-0.noarch
> >
> > ceph-release-1-0.el6.noarch
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> >
> > Here's rpm -qa from the mons:
> >
> >
> > [root@ceph-mon01 ~]# rpm -qa|grep ceph
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > ceph-release-1-0.el6.noarch
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> > [root@ceph-mon01 ~]#
> >
> >
> > [root@ceph-mon02 ~]# rpm -qa|grep ceph
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > ceph-release-1-0.el6.noarch
> >
> > [root@ceph-mon02 ~]#
> >
> >
> > [root@ceph-mon03 ~]# rpm -qa|grep ceph
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > ceph-release-1-0.el6.noarch
> >
> > [root@ceph-mon03 ~]#
> >
> >
> > Joshua
> >
> >
> >
> > On Thu, Jul 10, 2014 at 6:04 PM, Sage Weil <[email protected]> wrote:
> > Have you made any other changes after the upgrade? (Like
> > adjusting
> > tunables, or creating EC pools?)
> >
> > See if there is anything in 'dmesg' output.
> >
> > sage
> >
> > On Thu, 10 Jul 2014, Joshua McClintock wrote:
> >
> > > I upgraded my cluster to .80.1-2 (CentOS). My mount command
> > just freezes
> > > and outputs an error:
> > >
> > > mount.ceph 192.168.0.14,192.168.0.15,192.168.0.16:/ /us-west01
> > -o
> > > name=chefwks01,secret=`ceph-authtool -p -n client.admin
> > > /etc/ceph/us-west01.client.admin.keyring`
> > >
> > > mount error 5 = Input/output error
> > >
> > >
> > > Here's the output from 'ceph -s'
> > >
> > >
> > > cluster xxxxxxxxxxxxxxxxxxxxxx
> > >
> > > health HEALTH_OK
> > >
> > > monmap e1: 3 monsat{ceph-mon01=
> 192.168.0.14:6789/0,ceph-mon02=192.168.0.15:6789/0,ceph-mon03
> > =1
> > > 92.168.0.16:6789/0}, election epoch 88, quorum 0,1,2
> > > ceph-mon01,ceph-mon02,ceph-mon03
> > >
> > > mdsmap e26: 1/1/1 up {0=0=up:active}
> > >
> > > osdmap e1371: 5 osds: 5 up, 5 in
> > >
> > > pgmap v49431: 192 pgs, 3 pools, 135 GB data, 34733
> > objects
> > >
> > > 406 GB used, 1874 GB / 2281 GB avail
> > >
> > > 192 active+clean
> > >
> > >
> > > I can see some packets being exchanged between the client and
> > the mon, but
> > > it's a pretty short exchange.
> > >
> > > Any ideas where to look next?
> > >
> > > Joshua
> > >
> > >
> > >
> >
> >
> >
> >
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com