We use 3 servers to build a test system of ceph, configured as below:
Host IP
client01 192.168.1.10
ceph01 192.168.2.50
ceph02 192.168.2.51
The OS is unbuntu 10.04 LTS and the version of ceph is v0.21.1
ceph.conf:
[global]
auth supported = cephx
pid file = /var/run/ceph/$name.pid
debug ms = 0
keyring = /etc/ceph/keyring.bin
[mon]
mon data = /mnt/ceph/data/mon$id
debug ms = 1
[mon0]
host = ceph01
mon addr = 192.168.2.50:6789
[mds]
keyring = /etc/ceph/keyring.$name
debug ms = 1
[mds.ceph01]
host = ceph01
[mds.ceph02]
host = ceph02
[osd]
sudo = true
osd data = /mnt/ceph/osd$id/data
keyring = /etc/ceph/keyring.$name
osd journal = /mnt/ceph/osd$id/data/journal
osd journal size = 100
[osd0]
host = ceph01
[osd1]
host = ceph01
[osd2]
host = ceph01
[osd3]
host = ceph01
[osd10]
host = ceph02
There are 4 HDDs in the ceph01 and every HDD has a OSD named as osd0, osd1,
osd2,osd3; there is 1 HDD in the ceph02 named as osd10. All these HDDs are made
as btrfs and mounted on the mount point as listed below:
ceph01
/dev/sdc1 /mnt/ceph/osd0/data btrfs
/dev/sdd1 /mnt/ceph/osd1/data btrfs
/dev/sde1 /mnt/ceph/osd2/data btrfs
/dev/sdf1 /mnt/ceph/osd3/data btrfs
ceph02
/dev/sdb1 /mnt.ceph/osd10/data btrfs
Make ceph FileSystem:
r...@ceph01:~# mkcephfs -c /etc/cepf/ceph.conf -a -k /etc/ceph/keyring.bin
Startup ceph:
r...@ceph01:~# /etc/init.d/ceph –a start
Then
r...@ceph01:~# ceph -w
10.09.01_17:56:19.337895 mds e17: 1/1/1 up {0=up:active}, 1 up:standby
10.09.01_17:56:19.347184 osd e27: 5 osds: 5 up, 5 in
10.09.01_17:56:19.349447 log …
10.09.01_17:56:19.373773 mon e1: 1 mons at 192.168.2.50:6789/0
The ceph file system is mounted to client01(192.168.1.10),
ceph01(192.168.2.50), ceph02(192.168.2.51)at /data/ceph. It works fine at the
beginning, I can use ls and the write and read of file is ok. After some files
are wrote , I find I can’t use ls –l /data/ceph until I umount ceph from
ceph02, but one day later the same problem occurred again, then I umount ceph
from ceph01 the system and everything is ok.
Q1:
Can the ceph filesystem be mounted to a member of ceph cluster?
When I follow the instruction of
http://ceph.newdream.net/wiki/Monitor_cluster_expansion to expand a monitor to
ceph02, the following error occurred:
r...@ceph02:~# /etc/init.d/ceph start mon1
[/etc/ceph/fetch_config/tmp/fetched.ceph.conf.14210] ceph.conf 100% 2565
2.5KB/s 00:00
=== mon.1 ===
Starting Ceph mon1 on ceph02...
** WARNING: Ceph is still under heavy development, and is only suitable for
**
** testing and review. Do not trust it with important data. **
terminate called after throwing an instance of 'std::logic_error'
what(): basic_string::_S_construct NULL not valid
Aborted (core dumped)
failed: ' /usr/bin/cmon -i 1 -c /tmp/fetched.ceph.conf.14210 '
Q2:
How to expand a monitor to a running ceph system?
Q3
Is it possible to add mds when the ceph system is running? how?
I fdisked a HDD into two partition, one for journal, other one for data like
this:
/dev/sdc1(180GB)as data
/dev/sdc2(10GB)as journal
/dev/sdc1 made as btrfs, mount to /mnt/osd0/data
/dev/sdc2 made as btrfs, mount to /mnt/osd0/journal
ceph.conf:
…
[osd]
osd data = /mnt/ceph/osd$id/data
osd journal = /mnt/ceph/osd$id/journal
; osd journal size = 100
…
When I use mkcephfs command, I can't build osd until I edited ceph.conf like
this:
[osd]
osd data = /mnt/ceph/osd$id/data
osd journal = /mnt/ceph/osd$id/data/journal
osd journal size = 100
…
Q4.
How to set the journal path to a device or patition?
Thanks for all help and reply , sorry for my lame English.
Lin
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html