G'day all
Today's testing was on having a client (i.e. Ubuntu 12.04 server running KVM to
be virtualisation host, named "server") connect to the 3x node v0.42 Ceph
Cluster (names "ceph1, ceph2 & ceph3").
A very simple command "rbd ls" is failing per below:
root@server:/# rbd ls
terminate called after throwing an instance of 'ceph::buffer::error'
what(): buffer::exception
*** Caught signal (Aborted) **
in thread 7fe239f7d700
ceph version 0.41 (commit:c1345f7136a0af55d88280ffe4b58339aaf28c9d)
1: rbd() [0x459316]
2: (()+0xfcb0) [0x7fe23c055cb0]
3: (gsignal()+0x35) [0x7fe23a7b5475]
4: (abort()+0x17b) [0x7fe23a7b8bdb]
5: (__gnu_cxx::__verbose_terminate_handler()+0x11d) [0x7fe23b10169d]
6: (()+0xb5846) [0x7fe23b0ff846]
7: (()+0xb5873) [0x7fe23b0ff873]
8: (()+0xb596e) [0x7fe23b0ff96e]
9: (pg_pool_t::decode(ceph::buffer::list::iterator&)+0x859) [0x7fe23c3d0e99]
10: (void decode<long, pg_pool_t>(std::map<long, pg_pool_t, std::less<long>,
std::allocator<std::pair<long const, pg_pool_t> > >&,
ceph::buffer::list::iterator&)+0x91) [0x7fe23c3c8851]
11: (OSDMap::decode(ceph::buffer::list&)+0x17d) [0x7fe23c3befcd]
12: (Objecter::handle_osd_map(MOSDMap*)+0x1e9f) [0x7fe23c3033ff]
13: (librados::RadosClient::_dispatch(Message*)+0x54) [0x7fe23c2e3e44]
14: (librados::RadosClient::ms_dispatch(Message*)+0x33) [0x7fe23c2e3ef3]
15: (SimpleMessenger::dispatch_entry()+0x76b) [0x47ea1b]
16: (SimpleMessenger::DispatchThread::entry()+0x1c) [0x46995c]
17: (()+0x7e9a) [0x7fe23c04de9a]
18: (clone()+0x6d) [0x7fe23a86f77d]
Aborted (core dumped)
However this command works on the same box:
root@server:/# echo "10.112.9.11,10.112.9.12,10.112.9.13
name=admin,secret=AQBBg0NPeFt9MRAAg8AWG108I8sje82hzdjkiw== rbd demo.rbd" >
/sys/bus/rbd/add
root@server:/# ls -alh /dev/rbd*
brw-rw---- 1 root disk 251, 0 Feb 22 14:12 /dev/rbd0
/dev/rbd:
total 0
drwxr-xr-x 3 root root 60 Feb 22 14:12 .
drwxr-xr-x 16 root root 4.2K Feb 22 14:12 ..
drwxr-xr-x 2 root root 60 Feb 22 14:12 rbd
Confirming the below works as expected when run from any of the 3x servers
within the Cluster:
root@ceph1:/etc/ceph# rbd ls
demo.rbd
root@ceph3:~# rbd -p rbd ls
demo.rbd
On the "client" KVM box, I have installed:
apt-get install ceph-common
modprobe rbd
Then created these two files:
;-------------------------------------------------------------------
/etc/ceph/ceph.conf
;-------------------------------------------------------------------
[global]
; enable secure authentication
auth supported = cephx
keyring = /etc/ceph/keyring.bin
[mon]
[mon.alpha]
host = ceph1
mon addr = 10.112.9.11:6789
[mon.bravo]
host = ceph2
mon addr = 10.112.9.12:6789
[mon.charlie]
host = ceph3
mon addr = 10.112.9.13:6789
;-------------------------------------------------------------------
/etc/ceph/keyring.bin
;-------------------------------------------------------------------
[client.admin]
key = AQBBg0NPeFt9MRAAg8AWG108I8sje82hzdjkiw==
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html