Re: [ceph-users] Compile from source with Kinetic support

2014-12-17 Thread Julien Lutran

Hello,

Ceph is still not compiling when if I add Kinetic support option.
Could you have a look on the log and tell me what's missing ?

--
Best regards,
Julien

On 12/02/2014 09:53 AM, Julien Lutran wrote:
It's ok for KeyValueDB.cc now, but I have another problem with 
src/os/KineticStore.h :


http://pastebin.com/nhSDQs70


On 12/01/2014 05:58 PM, Haomai Wang wrote:

Sorry, it's a typo

/WITH_KINETIC/HAVE_KINETIC/

:-)

On Tue, Dec 2, 2014 at 12:51 AM, Julien Lutran julien.lut...@ovh.net 
mailto:julien.lut...@ovh.net wrote:



Sorry, It didn't change anything :

root@host:~/sources/ceph# head -12 src/os/KeyValueDB.cc
// -*- mode:C++; tab-width:8; c-basic-offset:2;
indent-tabs-mode:t -*-
// vim: ts=8 sw=2 smarttab

#include KeyValueDB.h
#include LevelDBStore.h
#ifdef HAVE_LIBROCKSDB
#include RocksDBStore.h
#endif
#ifdef WITH_KINETIC
#include KineticStore.h
#endif

root@host:~/sources/ceph# make
[...]
  CXX  os/libos_la-KeyValueDB.lo
os/KeyValueDB.cc: In static member function 'static KeyValueDB*
KeyValueDB::create(CephContext*, const string, const string)':
os/KeyValueDB.cc:21:16: error: expected type-specifier before
'KineticStore'
 return new KineticStore(cct);
^
os/KeyValueDB.cc:21:16: error: expected ';' before 'KineticStore'
os/KeyValueDB.cc:21:32: error: 'KineticStore' was not declared in
this scope
 return new KineticStore(cct);
^
os/KeyValueDB.cc: In static member function 'static int
KeyValueDB::test_init(const string, const string)':
os/KeyValueDB.cc:39:12: error: 'KineticStore' has not been declared
 return KineticStore::_test_init(g_ceph_context);
^
make[3]: *** [os/libos_la-KeyValueDB.lo] Error 1


On 12/01/2014 03:22 PM, Haomai Wang wrote:

#ifdef WITH_KINETIC
#include KineticStore.h
#endif





--

Best Regards,

Wheat





___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] 'rbd list' stuck

2014-12-17 Thread yang . bin18
Hi

Why command 'rbd list'executed on monitor stuck,any prompt should be 
appreciated!

Backtree: 


[810bfdee] futex_wait_queue_me+0xde/0x140
[810c0969] futex_wait+0x179/0x280
[810c297e] do_futex+0xfe/0x5e0
[810c2ee0] SyS_futex+0x80/0x180
[815f2119] system_call_fastpath+0x16/0x1b
[] 0x

Best Regards!

 YangBin

ZTE Information Security Notice: The information contained in this mail (and 
any attachment transmitted herewith) is privileged and confidential and is 
intended for the exclusive use of the addressee(s).  If you are not an intended 
recipient, any disclosure, reproduction, distribution or other dissemination or 
use of the information contained is strictly prohibited.  If you have received 
this mail in error, please delete it and notify us immediately.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Help with Integrating Ceph with various Cloud Storage

2014-12-17 Thread Karan Singh
Hello Manoj

My answers to your queries.

# For testing purpose you can install Ceph on virtual machines ( multiple 
instances of virtual box for multiple mon, osd ). Its good to practice Ceph 
with multiple MON and OSD.
# For real data storage , please use physical servers , virtual servers are 
good to go only for testing and learning purpose.

# S3,azure blob,swift are object storage which are different from each other , 
similarly Ceph also supports object storage and you can access Ceph object 
storage with S3 and Swift protocols. You cannot use Ceph storage with 
amazon,azure but you can use ceph with Swift and s3 API.

# I am not much familiar with cloud storage , but based on the experience of 
openstack , libvirt should be installed from cloudstack servers only , Ceph 
servers should only do storage for you.


Karan Singh 
Systems Specialist , Storage Platforms
CSC - IT Center for Science,
Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
mobile: +358 503 812758
tel. +358 9 4572001
fax +358 9 4572302
http://www.csc.fi/


On 17 Dec 2014, at 09:57, Manoj Singh respond2ma...@gmail.com wrote:

 Hi All,
 
 I am new to Ceph. Due to physical machines shortage I have installed Ceph 
 cluster with single OSD and MON in a single Virtual Machine.  
 
 I have few queries as below:
 1.  Whether having the Ceph setup on a VM is fine or it require to be on 
 Physical server.
 
 2. Since Amazon S3, Azure Blob Storage, Swift are Object based Storage, what 
 is the feasibility of attaching these Cloud Storage to Ceph and to be able to 
 allocate disc space from the same while creating new VM from local CloudStack 
 or OpenStack
 
 3. When I am integrating CloudStack with Ceph whether libvert should be 
 installed on the CloudStack management server or on Ceph server. From diagram 
 given in Ceph documentation it's bit confusing.
 
 Thank you in advance. your help shall be really appreciated.
 
 Best Regards,
 Manoj Kumar
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



smime.p7s
Description: S/MIME cryptographic signature
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Is cache tiering production ready?

2014-12-17 Thread Yujian Peng
Hi,
Since firefly, ceph can support cache tiering.
Cache tiering: support for creating ‘cache pools’ that store hot, recently
accessed objects with automatic demotion of colder data to a base tier.
Typically the cache pool is backed by faster storage devices like SSDs.

I'm testing cache tiering, and everything goes well.

But rados -h shows that:
CACHE POOLS: (for testing/development only)
   cache-flush obj-name   flush cache pool object (blocking)
   cache-try-flush obj-name   flush cache pool object (non-blocking)
   cache-evict obj-name   evict cache pool object
   cache-flush-evict-allflush+evict all objects
   cache-try-flush-evict-alltry-flush+evict all objects

rados -v
ceph version 0.80.7 (6c0127fcb58008793d3c8b62d925bc91963672a3)

Is cache tiering production ready?

Thanks a lot!

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] File System stripping data

2014-12-17 Thread Wido den Hollander
On 12/17/2014 11:21 AM, John Spray wrote:
 On Wed, Dec 17, 2014 at 2:07 AM, Kevin Shiah agan...@gmail.com wrote:
 setfattr -n ceph.dir.layout.stripe_count -v 2 dir

 And return:

 setfattr: dir: Operation not supported
 
 Works for me on master.  What ceph version are you using?
 

I just tried something similar on Giant (0.87) and I saw this in the logs:

parse_layout_vxattr name layout.pool value 'cephfs_svo'
 invalid data pool 3
reply request -22

I resolves the pool to a ID, but then it's unable to set it?

 John
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
 


-- 
Wido den Hollander
42on B.V.
Ceph trainer and consultant

Phone: +31 (0)20 700 9902
Skype: contact42on
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] File System stripping data

2014-12-17 Thread Wido den Hollander
On 12/17/2014 12:35 PM, John Spray wrote:
 On Wed, Dec 17, 2014 at 10:25 AM, Wido den Hollander w...@42on.com wrote:
 I just tried something similar on Giant (0.87) and I saw this in the logs:

 parse_layout_vxattr name layout.pool value 'cephfs_svo'
  invalid data pool 3
 reply request -22

 I resolves the pool to a ID, but then it's unable to set it?
 
 Was the 'cephfs_svo' pool already added as a data pool with ceph mds
 add_data_pool?
 

Ah, indeed. Working fine right now. Same goes for any other layout settings.

 There are paths where if a pool was added very recently, MDSs/clients
 might not know about the pool yet and can generate errors like this.
 
 John
 


-- 
Wido den Hollander
42on B.V.
Ceph trainer and consultant

Phone: +31 (0)20 700 9902
Skype: contact42on
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Italo Santos
Hello Loic,  

Thanks for you help, I’ve take a look to my crush map and I replace step 
chooseleaf indep 0 type osd” by step choose indep 0 type osd” and all PGs was 
created successfully.  

At.

Italo Santos
http://italosantos.com.br/


On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:

 Hi,
  
 The 2147483647 means that CRUSH did not find enough OSD for a given PG. If 
 you check the crush rule associated with the erasure coded pool, you will 
 most probably find why.
  
 Cheers
  
 On 16/12/2014 23:32, Italo Santos wrote:
  Hello,
   
  I'm trying to create an erasure pool following 
  http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but when I 
  try create a pool with a specifc erasure-code-profile (myprofile) the PGs 
  became on incomplete state.
   
  Anyone can help me?
   
  Below the profile I created:
  root@ceph0001:~# ceph osd erasure-code-profile get myprofile
  directory=/usr/lib/ceph/erasure-code
  k=6
  m=2
  plugin=jerasure
  technique=reed_sol_van
   
  The status of cluster:
  root@ceph0001:~# ceph health
  HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean
   
  health detail:
  root@ceph0001:~# ceph health detail
  HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean
  pg 2.9 is stuck inactive since forever, current state incomplete, last 
  acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
  pg 2.8 is stuck inactive since forever, current state incomplete, last 
  acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
  pg 2.b is stuck inactive since forever, current state incomplete, last 
  acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
  pg 2.a is stuck inactive since forever, current state incomplete, last 
  acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
  pg 2.5 is stuck inactive since forever, current state incomplete, last 
  acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
  pg 2.4 is stuck inactive since forever, current state incomplete, last 
  acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
  pg 2.7 is stuck inactive since forever, current state incomplete, last 
  acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
  pg 2.6 is stuck inactive since forever, current state incomplete, last 
  acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
  pg 2.1 is stuck inactive since forever, current state incomplete, last 
  acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
  pg 2.0 is stuck inactive since forever, current state incomplete, last 
  acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
  pg 2.3 is stuck inactive since forever, current state incomplete, last 
  acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
  pg 2.2 is stuck inactive since forever, current state incomplete, last 
  acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
  pg 2.9 is stuck unclean since forever, current state incomplete, last 
  acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
  pg 2.8 is stuck unclean since forever, current state incomplete, last 
  acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
  pg 2.b is stuck unclean since forever, current state incomplete, last 
  acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
  pg 2.a is stuck unclean since forever, current state incomplete, last 
  acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
  pg 2.5 is stuck unclean since forever, current state incomplete, last 
  acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
  pg 2.4 is stuck unclean since forever, current state incomplete, last 
  acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
  pg 2.7 is stuck unclean since forever, current state incomplete, last 
  acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
  pg 2.6 is stuck unclean since forever, current state incomplete, last 
  acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
  pg 2.1 is stuck unclean since forever, current state incomplete, last 
  acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
  pg 2.0 is stuck unclean since forever, current state incomplete, last 
  acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
  pg 2.3 is stuck unclean since forever, current state incomplete, last 
  acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
  pg 2.2 is stuck unclean since forever, current state incomplete, last 
  acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
  pg 2.9 is incomplete, acting 
  [4,10,15,2147483647,3,2147483647,2147483647,2147483647] (reducing pool 
  ecpool min_size from 6 may help; search ceph.com/docs 
  (http://ceph.com/docs) for 'incomplete')
  pg 2.8 is incomplete, acting 
  [0,2147483647,4,2147483647,10,2147483647,15,2147483647] (reducing pool 
  ecpool min_size from 6 may help; search ceph.com/docs 
  (http://ceph.com/docs) for 

[ceph-users] cephfs not mounting on boot

2014-12-17 Thread Lindsay Mathieson
Both fuse and kernel module fail to mount,

The mons  mds are on two other nodes, so they are available when this node is 
booting.

They can be mounted manually after boot.

my fstab:

  id­min  /mnt/cephfs  fuse.ceph defaults,nonempty,_netdev 0 0
  vnb.proxmox.softlog,vng.proxmox.softlog,vnt.proxmox.softlog:/   /mnt/test 
 ceph_netdev,defaults,name­min,secretfile=/etc/pve/priv/admin.secret 0 0
--
Lindsay


signature.asc
Description: This is a digitally signed message part.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs kernel module reports error on mount

2014-12-17 Thread John Spray
Hmm, from a quick google it appears you are not the only one who has
seen this symptom with mount.ceph.  Our mtab code appears to have
diverged a bit from the upstream util-linux repo, so it seems entirely
possible we have a bug in ours somewhere.  I've opened
http://tracker.ceph.com/issues/10351 to track it.

Cheers,
John

On Wed, Dec 17, 2014 at 1:31 PM, Lindsay Mathieson
lindsay.mathie...@gmail.com wrote:
 mount reports:

 mount: error writing /etc/mtab: Invalid argument



 fstab entry is:



 vnb.proxmox.softlog,vng.proxmox.softlog,vnt.proxmox.softlog:/ /mnt/test ceph
 _netdev,defaults,name=admin,secretfile=/etc/pve/priv/admin.secret 0 0





 However the mounts is successful and a mtab entry is made.



 debian wheezy, ceph 0.87



 --

 Lindsay


 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs not mounting on boot

2014-12-17 Thread John Spray
Can you tell us more about how they fail?  Error messages on console,
anything in syslog?

In the absence of other clues, you might want to try checking that the
network is coming up before ceph tries to mount.

John

On Wed, Dec 17, 2014 at 1:34 PM, Lindsay Mathieson
lindsay.mathie...@gmail.com wrote:
 Both fuse and kernel module fail to mount,



 The mons  mds are on two other nodes, so they are available when this node
 is booting.



 They can be mounted manually after boot.



 my fstab:



 id=admin /mnt/cephfs fuse.ceph defaults,nonempty,_netdev 0 0

 vnb.proxmox.softlog,vng.proxmox.softlog,vnt.proxmox.softlog:/ /mnt/test ceph
 _netdev,defaults,name=admin,secretfile=/etc/pve/priv/admin.secret 0 0

 --

 Lindsay


 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs not mounting on boot

2014-12-17 Thread Lindsay Mathieson
On Wed, 17 Dec 2014 02:02:52 PM John Spray wrote:
 Can you tell us more about how they fail?  Error messages on console,
 anything in syslog?

Not quite sure what to look for, but I did a quick scan on ceph through dmesg 
 syslog, nothing stood out


 
 In the absence of other clues, you might want to try checking that the
 network is coming up before ceph tries to mount.

Now I think on it, that might just be it - I seem to recall a similar problem 
with cifs mounts, despite having the _netdev option. I had to issue a mount in  
/etc/network/if-up.d/

I'll test than and get back to you

-- 
Lindsay

signature.asc
Description: This is a digitally signed message part.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Is cache tiering production ready?

2014-12-17 Thread Gregory Farnum
Cache tiering is a stable, functioning system. Those particular commands
are for testing and development purposes, not something you should run
(although they ought to be safe).
-Greg
On Wed, Dec 17, 2014 at 1:44 AM Yujian Peng pengyujian5201...@126.com
wrote:

 Hi,
 Since firefly, ceph can support cache tiering.
 Cache tiering: support for creating ‘cache pools’ that store hot, recently
 accessed objects with automatic demotion of colder data to a base tier.
 Typically the cache pool is backed by faster storage devices like SSDs.

 I'm testing cache tiering, and everything goes well.

 But rados -h shows that:
 CACHE POOLS: (for testing/development only)
cache-flush obj-name   flush cache pool object (blocking)
cache-try-flush obj-name   flush cache pool object (non-blocking)
cache-evict obj-name   evict cache pool object
cache-flush-evict-allflush+evict all objects
cache-try-flush-evict-alltry-flush+evict all objects

 rados -v
 ceph version 0.80.7 (6c0127fcb58008793d3c8b62d925bc91963672a3)

 Is cache tiering production ready?

 Thanks a lot!

 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Ceph rbd mapped but files all have 0 byte size

2014-12-17 Thread Cyan Cheng
Dear All,

We have set up ceph and used it for about one year already. 

Here is a summary of the setting. We used 3 servers to run the ceph.

cs02, cs03, cs04

Here is how we set up the ceph:

1. We created several OSDs on three of these servers. using command like:

 ceph-deploy osd create cs02:/dev/sdc …. cs03:/dev/… cs04:/dev/….

2. And have created MDS on cs02:

 ceph-deploy mds create ilab-cs02

3. After that, we have created a RADOS block device on cs02 by 

 rbd create rbd-research --size 1024

4. Then mapped rbd-research

 sudo tbd map rbd-resrearch —pool rbd

5. Then make file system 

 sudo mkfs.ext4 /dev/rbd/rbd/rbd-research

6. Then mkdir and mount the rbd by adding this line to /etc/fstab

/dev/rbd/rbd/rbd-research /mnt/retinadata ext4 defaults,users 0 2

7. Then mount

 mount /mnt/retinadata

It worked reliably until recently we had a power off of our servers 
accidentally.

After power recovered. cs03, cs04 were automatically boot up, while cs02 were 
not automatically boot up. There is a message shown on cs02 telling something 
like “not able to mount /mnt/retinadata, device not found, press S to ignore 
and continue booting, press M to manual configure”. We selected S and booted up 
the system.

Then we found that, /mnt/retinadata was not mount and the rbd image at 
/dev/rbd/rbd/rbd1 was not there. 

We map the rbd image once again by.

 sudo tbd map rbd-research —pool rbd

Then we were able to mount /mnt/retinadata 

But the result we have now is :
1. All the file system structures are there.
2. All the files are of 0 byte size.

Could anybody help on this issue? Thank you very much in advance.

Some more information. We tried to reboot cs02 again. And we see a full screen 
of error message like: 

[44038.215233] libceph: connect 192.168.1.31:6789 socket error on write
[44038.215308] libceph: mon1 192.1.168.31:6789 error -101
libceph: connect 192.168.1.41:6812 error -101
libceph: osd22 192.168.1.41:6812 socket error on write 

Best Regards,

Cyan Cheng






___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Help with SSDs

2014-12-17 Thread Bryson McCutcheon
Hey there,

Is there a good work around if our SSDs are not handling D_SYNC very well? We 
invested a ton of money into Samsung 840 EVOS and they are not playing well 
with D_SYNC.  Would really appreciate the help!

Thank you,

Bryson



___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Incorrect description in document at chapter 'Crush Operation'?

2014-12-17 Thread 童磊

Hi,all
  I found content below at 
http://ceph.com/docs/master/rados/operations/crush-map :
 
  step choose firstn {num} type {bucket-type}
Description: Selects the number of buckets of the given type. The 
number is usually the number of replicas in the pool (i.e., pool size).
If {num} == 0, choose pool-num-replicas buckets (all available).
If {num}  0   pool-num-replicas, choose that many buckets.
If {num}  0, it means pool-num-replicas - {num}.
  
  However, corresponding source code is that (V0.87, mapper.c , 
crush_do_rule(...)) :
if (numrep = 0) {
  numrep += result_max;
  if (numrep = 0)
continue;
}

  I think the correct description is 'If {num}  0, it means pool-num-replicas 
+ {num}'.

--
TongLei | XOR Media

Storage Application Department of XOR Media China

Tel. 86-21-61213323  Fax. 86-21-52065237
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph rbd mapped but files all have 0 byte size

2014-12-17 Thread Ilya Dryomov
On Tue, Dec 16, 2014 at 6:19 AM, Cyan Cheng cheng.1...@gmail.com wrote:
 Dear All,

 We have set up ceph and used it for about one year already.

 Here is a summary of the setting. We used 3 servers to run the ceph.

 cs02, cs03, cs04

 Here is how we set up the ceph:

 1. We created several OSDs on three of these servers. using command like:

 ceph-deploy osd create cs02:/dev/sdc …. cs03:/dev/… cs04:/dev/….

 2. And have created MDS on cs02:

 ceph-deploy mds create ilab-cs02

 3. After that, we have created a RADOS block device on cs02 by

 rbd create rbd-research --size 1024

 4. Then mapped rbd-research

 sudo tbd map rbd-resrearch —pool rbd

I assume that's rbd map.  Mapping images on the same physical box
that's also running OSDs works in general but isn't a very good idea.


 5. Then make file system

 sudo mkfs.ext4 /dev/rbd/rbd/rbd-research

 6. Then mkdir and mount the rbd by adding this line to /etc/fstab

 /dev/rbd/rbd/rbd-research /mnt/retinadata ext4 defaults,users 0 2

 7. Then mount

 mount /mnt/retinadata

 It worked reliably until recently we had a power off of our servers
 accidentally.

 After power recovered. cs03, cs04 were automatically boot up, while cs02
 were not automatically boot up. There is a message shown on cs02 telling
 something like “not able to mount /mnt/retinadata, device not found, press S
 to ignore and continue booting, press M to manual configure”. We selected S
 and booted up the system.

 Then we found that, /mnt/retinadata was not mount and the rbd image at
 /dev/rbd/rbd/rbd1 was not there.

 We map the rbd image once again by.

 sudo tbd map rbd-research —pool rbd

 Then we were able to mount /mnt/retinadata

 But the result we have now is :
 1. All the file system structures are there.
 2. All the files are of 0 byte size.

All files or just those that you (your system) were working with at the
time of the power reset?


 Could anybody help on this issue? Thank you very much in advance.

 Some more information. We tried to reboot cs02 again. And we see a full
 screen of error message like:

 [44038.215233] libceph: connect 192.168.1.31:6789 socket error on write
 [44038.215308] libceph: mon1 192.1.168.31:6789 error -101
 libceph: connect 192.168.1.41:6812 error -101
 libceph: osd22 192.168.1.41:6812 socket error on write

That's Network is unreachable so probably something's wrong with your
network.

Thanks,

Ilya
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Certificate has expired

2014-12-17 Thread Emilio

Hi,

We have some problems with ceph-deploy install node

This is the error I get when I run the installation:

[mon01][INFO  ] Running command: sudo rpm --import 
https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc
[mon01][INFO  ] Running command: sudo rpm --import 
https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc
[mon01][WARNIN] curl: (60) Peer's Certificate has expired.











Can you fix this problem? We can run the install with other parameters?

Thanks!
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Certificate has expired

2014-12-17 Thread John Spray
Strange, when I visit https://ceph.com, I get a certificate that
doesn't expire until 10 February 2015.

Perhaps check the clock on your node isn't in the future?

John

On Wed, Dec 17, 2014 at 4:16 PM, Emilio emilio.mor...@adam.es wrote:
 Hi,

 We have some problems with ceph-deploy install node

 This is the error I get when I run the installation:

 [mon01][INFO  ] Running command: sudo rpm --import
 https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc
 [mon01][INFO  ] Running command: sudo rpm --import
 https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc
 [mon01][WARNIN] curl: (60) Peer's Certificate has expired.











 Can you fix this problem? We can run the install with other parameters?

 Thanks!

 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Loic Dachary
Hi,

Thanks for the update : good news are much appreciated :-) Would you have time 
to review the documentation at https://github.com/ceph/ceph/pull/3194/files ? 
It was partly motivated by the problem you had.

Cheers

On 17/12/2014 14:03, Italo Santos wrote:
 Hello Loic,
 
 Thanks for you help, I’ve take a look to my crush map and I replace step 
 chooseleaf indep 0 type osd” by step choose indep 0 type osd” and all PGs 
 was created successfully.
 
 At.
 
 *Italo Santos*
 http://italosantos.com.br/
 
 On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
 
 Hi,

 The 2147483647 means that CRUSH did not find enough OSD for a given PG. If 
 you check the crush rule associated with the erasure coded pool, you will 
 most probably find why.

 Cheers

 On 16/12/2014 23:32, Italo Santos wrote:
 Hello,

 I'm trying to create an erasure pool following 
 http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but when I 
 try create a pool with a specifc erasure-code-profile (myprofile) the PGs 
 became on incomplete state.

 Anyone can help me?

 Below the profile I created:
 root@ceph0001:~# ceph osd erasure-code-profile get myprofile
 directory=/usr/lib/ceph/erasure-code
 k=6
 m=2
 plugin=jerasure
 technique=reed_sol_van

 The status of cluster:
 root@ceph0001:~# ceph health
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean

 health detail:
 root@ceph0001:~# ceph health detail
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean
 pg 2.9 is stuck inactive since forever, current state incomplete, last 
 acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck inactive since forever, current state incomplete, last 
 acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck inactive since forever, current state incomplete, last 
 acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck inactive since forever, current state incomplete, last 
 acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
 pg 2.5 is stuck inactive since forever, current state incomplete, last 
 acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
 pg 2.4 is stuck inactive since forever, current state incomplete, last 
 acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
 pg 2.7 is stuck inactive since forever, current state incomplete, last 
 acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
 pg 2.6 is stuck inactive since forever, current state incomplete, last 
 acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
 pg 2.1 is stuck inactive since forever, current state incomplete, last 
 acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
 pg 2.0 is stuck inactive since forever, current state incomplete, last 
 acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
 pg 2.3 is stuck inactive since forever, current state incomplete, last 
 acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
 pg 2.2 is stuck inactive since forever, current state incomplete, last 
 acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
 pg 2.9 is stuck unclean since forever, current state incomplete, last 
 acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck unclean since forever, current state incomplete, last 
 acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck unclean since forever, current state incomplete, last 
 acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck unclean since forever, current state incomplete, last 
 acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
 pg 2.5 is stuck unclean since forever, current state incomplete, last 
 acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
 pg 2.4 is stuck unclean since forever, current state incomplete, last 
 acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
 pg 2.7 is stuck unclean since forever, current state incomplete, last 
 acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
 pg 2.6 is stuck unclean since forever, current state incomplete, last 
 acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
 pg 2.1 is stuck unclean since forever, current state incomplete, last 
 acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
 pg 2.0 is stuck unclean since forever, current state incomplete, last 
 acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
 pg 2.3 is stuck unclean since forever, current state incomplete, last 
 acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
 pg 2.2 is stuck unclean since forever, current state incomplete, last 
 acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
 pg 2.9 is incomplete, acting 
 [4,10,15,2147483647,3,2147483647,2147483647,2147483647] (reducing pool 
 ecpool min_size from 6 may help; search ceph.com/docs 
 http://ceph.com/docs for 'incomplete')
 pg 2.8 is 

Re: [ceph-users] Certificate has expired

2014-12-17 Thread Emilio

Yes, sorry this server was in the past!

Thx!


On 17/12/14 17:40, John Spray wrote:

Strange, when I visit https://ceph.com, I get a certificate that
doesn't expire until 10 February 2015.

Perhaps check the clock on your node isn't in the future?

John

On Wed, Dec 17, 2014 at 4:16 PM, Emilio emilio.mor...@adam.es wrote:

Hi,

We have some problems with ceph-deploy install node

This is the error I get when I run the installation:

[mon01][INFO  ] Running command: sudo rpm --import
https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc
[mon01][INFO  ] Running command: sudo rpm --import
https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc
[mon01][WARNIN] curl: (60) Peer's Certificate has expired.











Can you fix this problem? We can run the install with other parameters?

Thanks!

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




--
adamvozip.es Emilio Moreno Fernández
Departamento de Sistemas

www.adam.es La experiencia nos avala. 25 años de historia.
Tel. 902 902 685

Carrer Artesans, 7 - Parc Tecnològic del Vallès
08290 Cerdanyola del Vallès - Barcelona

www.adam.es http://www.adam.es


Advertencia legal:

Este mensaje y, en su caso, los ficheros anexos son confidenciales, 
especialmente en lo que respecta a los datos personales, y se dirigen 
exclusivamente al destinatario referenciado.
Si usted no lo es y lo ha recibido por error o tiene conocimiento del 
mismo por cualquier motivo, le rogamos que nos lo comunique por este 
medio y proceda a destruirlo o borrarlo, y que en todo caso se abstenga 
de utilizar, reproducir, alterar, archivar o comunicar a terceros el 
presente mensaje y ficheros anexos, todo ello bajo pena de incurrir en 
responsabilidades legales. El emisor no garantiza la integridad, rapidez 
o seguridad del presente correo, ni se responsabiliza de 
posiblesperjuicios derivados de la captura, incorporaciones de virus o 
cualesquiera otras manipulaciones efectuadas por terceros.



ecotech No imprimas si no es necesario. Protejamos el Medio Ambiente.

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Italo Santos
Hello,  

I’ve take a look to this documentation (which help a lot) and if I understand 
right, when I set a profile like:

===
ceph osd erasure-code-profile set isilon k=8 m=2 ruleset-failure-domain=host
===

And create a pool following the recommendations on doc, I’ll need (100*16)/2 = 
800 PGs, I’ll need the sufficient number of hosts to support create total PGs?  

Regards.

Italo Santos
http://italosantos.com.br/


On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:

 Hi,
  
 Thanks for the update : good news are much appreciated :-) Would you have 
 time to review the documentation at 
 https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated by the 
 problem you had.
  
 Cheers
  
 On 17/12/2014 14:03, Italo Santos wrote:
  Hello Loic,
   
  Thanks for you help, I’ve take a look to my crush map and I replace step 
  chooseleaf indep 0 type osd” by step choose indep 0 type osd” and all PGs 
  was created successfully.
   
  At.
   
  *Italo Santos*
  http://italosantos.com.br/
   
  On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
   
   Hi,

   The 2147483647 means that CRUSH did not find enough OSD for a given PG. 
   If you check the crush rule associated with the erasure coded pool, you 
   will most probably find why.

   Cheers

   On 16/12/2014 23:32, Italo Santos wrote:
Hello,
 
I'm trying to create an erasure pool following 
http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but 
when I try create a pool with a specifc erasure-code-profile 
(myprofile) the PGs became on incomplete state.
 
Anyone can help me?
 
Below the profile I created:
root@ceph0001:~# ceph osd erasure-code-profile get myprofile
directory=/usr/lib/ceph/erasure-code
k=6
m=2
plugin=jerasure
technique=reed_sol_van
 
The status of cluster:
root@ceph0001:~# ceph health
HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
unclean
 
health detail:
root@ceph0001:~# ceph health detail
HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
unclean
pg 2.9 is stuck inactive since forever, current state incomplete, last 
acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
pg 2.8 is stuck inactive since forever, current state incomplete, last 
acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
pg 2.b is stuck inactive since forever, current state incomplete, last 
acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
pg 2.a is stuck inactive since forever, current state incomplete, last 
acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
pg 2.5 is stuck inactive since forever, current state incomplete, last 
acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
pg 2.4 is stuck inactive since forever, current state incomplete, last 
acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
pg 2.7 is stuck inactive since forever, current state incomplete, last 
acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
pg 2.6 is stuck inactive since forever, current state incomplete, last 
acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
pg 2.1 is stuck inactive since forever, current state incomplete, last 
acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
pg 2.0 is stuck inactive since forever, current state incomplete, last 
acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
pg 2.3 is stuck inactive since forever, current state incomplete, last 
acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
pg 2.2 is stuck inactive since forever, current state incomplete, last 
acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
pg 2.9 is stuck unclean since forever, current state incomplete, last 
acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
pg 2.8 is stuck unclean since forever, current state incomplete, last 
acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
pg 2.b is stuck unclean since forever, current state incomplete, last 
acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
pg 2.a is stuck unclean since forever, current state incomplete, last 
acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
pg 2.5 is stuck unclean since forever, current state incomplete, last 
acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
pg 2.4 is stuck unclean since forever, current state incomplete, last 
acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
pg 2.7 is stuck unclean since forever, current state incomplete, last 
acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
pg 2.6 is stuck unclean since forever, current state incomplete, last 
acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
 

Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Loic Dachary


On 17/12/2014 18:18, Italo Santos wrote:
 Hello,
 
 I’ve take a look to this documentation (which help a lot) and if I understand 
 right, when I set a profile like:
 
 ===
 ceph osd erasure-code-profile set isilon k=8 m=2 ruleset-failure-domain=host
 ===
 
 And create a pool following the recommendations on doc, I’ll need (100*16)/2 
 = 800 PGs, I’ll need the sufficient number of hosts to support create total 
 PGs?

You will need k+m = 10 host per OSD. If you only have 10 hosts that should be 
ok and the 800 PGs will use these 10 OSD in various orders. It also means that 
you will end up having 800 PG per OSD which is a bit too mche. If you have 20 
OSDs that will be better : each PG will get 10 OSD out of 20 and each OSD will 
have 400 PGs. Ideally you want the number of PG per OSD to be in the range 
(approximately) [20,300].

Cheers

 
 Regards.
 
 *Italo Santos*
 http://italosantos.com.br/
 
 On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:
 
 Hi,

 Thanks for the update : good news are much appreciated :-) Would you have 
 time to review the documentation at 
 https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated by 
 the problem you had.

 Cheers

 On 17/12/2014 14:03, Italo Santos wrote:
 Hello Loic,

 Thanks for you help, I’ve take a look to my crush map and I replace step 
 chooseleaf indep 0 type osd” by step choose indep 0 type osd” and all PGs 
 was created successfully.

 At.

 *Italo Santos*
 http://italosantos.com.br/

 On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:

 Hi,

 The 2147483647 means that CRUSH did not find enough OSD for a given PG. If 
 you check the crush rule associated with the erasure coded pool, you will 
 most probably find why.

 Cheers

 On 16/12/2014 23:32, Italo Santos wrote:
 Hello,

 I'm trying to create an erasure pool following 
 http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but when 
 I try create a pool with a specifc erasure-code-profile (myprofile) the 
 PGs became on incomplete state.

 Anyone can help me?

 Below the profile I created:
 root@ceph0001:~# ceph osd erasure-code-profile get myprofile
 directory=/usr/lib/ceph/erasure-code
 k=6
 m=2
 plugin=jerasure
 technique=reed_sol_van

 The status of cluster:
 root@ceph0001:~# ceph health
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean

 health detail:
 root@ceph0001:~# ceph health detail
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean
 pg 2.9 is stuck inactive since forever, current state incomplete, last 
 acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck inactive since forever, current state incomplete, last 
 acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck inactive since forever, current state incomplete, last 
 acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck inactive since forever, current state incomplete, last 
 acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
 pg 2.5 is stuck inactive since forever, current state incomplete, last 
 acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
 pg 2.4 is stuck inactive since forever, current state incomplete, last 
 acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
 pg 2.7 is stuck inactive since forever, current state incomplete, last 
 acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
 pg 2.6 is stuck inactive since forever, current state incomplete, last 
 acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
 pg 2.1 is stuck inactive since forever, current state incomplete, last 
 acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
 pg 2.0 is stuck inactive since forever, current state incomplete, last 
 acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
 pg 2.3 is stuck inactive since forever, current state incomplete, last 
 acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
 pg 2.2 is stuck inactive since forever, current state incomplete, last 
 acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
 pg 2.9 is stuck unclean since forever, current state incomplete, last 
 acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck unclean since forever, current state incomplete, last 
 acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck unclean since forever, current state incomplete, last 
 acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck unclean since forever, current state incomplete, last 
 acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
 pg 2.5 is stuck unclean since forever, current state incomplete, last 
 acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
 pg 2.4 is stuck unclean since forever, current state incomplete, last 
 acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
 pg 2.7 is stuck unclean since forever, 

Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Italo Santos
Loic,

So, if want have a failure domain by host, I’ll need set up a erasure profile 
which k+m = total number of hosts I have, right?  

Regards.

Italo Santos
http://italosantos.com.br/


On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:

  
  
 On 17/12/2014 18:18, Italo Santos wrote:
  Hello,
   
  I’ve take a look to this documentation (which help a lot) and if I 
  understand right, when I set a profile like:
   
  ===
  ceph osd erasure-code-profile set isilon k=8 m=2 ruleset-failure-domain=host
  ===
   
  And create a pool following the recommendations on doc, I’ll need 
  (100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to support 
  create total PGs?
  
 You will need k+m = 10 host per OSD. If you only have 10 hosts that should be 
 ok and the 800 PGs will use these 10 OSD in various orders. It also means 
 that you will end up having 800 PG per OSD which is a bit too mche. If you 
 have 20 OSDs that will be better : each PG will get 10 OSD out of 20 and each 
 OSD will have 400 PGs. Ideally you want the number of PG per OSD to be in the 
 range (approximately) [20,300].
  
 Cheers
  
   
  Regards.
   
  *Italo Santos*
  http://italosantos.com.br/
   
  On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:
   
   Hi,

   Thanks for the update : good news are much appreciated :-) Would you have 
   time to review the documentation at 
   https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated by 
   the problem you had.

   Cheers

   On 17/12/2014 14:03, Italo Santos wrote:
Hello Loic,
 
Thanks for you help, I’ve take a look to my crush map and I replace 
step chooseleaf indep 0 type osd” by step choose indep 0 type osd” 
and all PGs was created successfully.
 
At.
 
*Italo Santos*
http://italosantos.com.br/
 
On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
 
 Hi,
  
 The 2147483647 means that CRUSH did not find enough OSD for a given 
 PG. If you check the crush rule associated with the erasure coded 
 pool, you will most probably find why.
  
 Cheers
  
 On 16/12/2014 23:32, Italo Santos wrote:
  Hello,
   
  I'm trying to create an erasure pool following 
  http://docs.ceph.com/docs/master/rados/operations/erasure-code/, 
  but when I try create a pool with a specifc erasure-code-profile 
  (myprofile) the PGs became on incomplete state.
   
  Anyone can help me?
   
  Below the profile I created:
  root@ceph0001:~# ceph osd erasure-code-profile get myprofile
  directory=/usr/lib/ceph/erasure-code
  k=6
  m=2
  plugin=jerasure
  technique=reed_sol_van
   
  The status of cluster:
  root@ceph0001:~# ceph health
  HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
  unclean
   
  health detail:
  root@ceph0001:~# ceph health detail
  HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
  unclean
  pg 2.9 is stuck inactive since forever, current state incomplete, 
  last acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
  pg 2.8 is stuck inactive since forever, current state incomplete, 
  last acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
  pg 2.b is stuck inactive since forever, current state incomplete, 
  last acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
  pg 2.a is stuck inactive since forever, current state incomplete, 
  last acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
  pg 2.5 is stuck inactive since forever, current state incomplete, 
  last acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
  pg 2.4 is stuck inactive since forever, current state incomplete, 
  last acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
  pg 2.7 is stuck inactive since forever, current state incomplete, 
  last acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
  pg 2.6 is stuck inactive since forever, current state incomplete, 
  last acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
  pg 2.1 is stuck inactive since forever, current state incomplete, 
  last acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
  pg 2.0 is stuck inactive since forever, current state incomplete, 
  last acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
  pg 2.3 is stuck inactive since forever, current state incomplete, 
  last acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
  pg 2.2 is stuck inactive since forever, current state incomplete, 
  last acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
  pg 2.9 is stuck unclean since forever, current state incomplete, 
  last acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
  pg 2.8 is stuck unclean 

Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Loic Dachary


On 17/12/2014 19:22, Italo Santos wrote:
 Loic,
 
 So, if want have a failure domain by host, I’ll need set up a erasure profile 
 which k+m = total number of hosts I have, right?

Yes, k+m has to be = number of hosts.

 
 Regards.
 
 *Italo Santos*
 http://italosantos.com.br/
 
 On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:
 


 On 17/12/2014 18:18, Italo Santos wrote:
 Hello,

 I’ve take a look to this documentation (which help a lot) and if I 
 understand right, when I set a profile like:

 ===
 ceph osd erasure-code-profile set isilon k=8 m=2 ruleset-failure-domain=host
 ===

 And create a pool following the recommendations on doc, I’ll need 
 (100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to support 
 create total PGs?

 You will need k+m = 10 host per OSD. If you only have 10 hosts that should 
 be ok and the 800 PGs will use these 10 OSD in various orders. It also means 
 that you will end up having 800 PG per OSD which is a bit too mche. If you 
 have 20 OSDs that will be better : each PG will get 10 OSD out of 20 and 
 each OSD will have 400 PGs. Ideally you want the number of PG per OSD to be 
 in the range (approximately) [20,300].

 Cheers


 Regards.

 *Italo Santos*
 http://italosantos.com.br/

 On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:

 Hi,

 Thanks for the update : good news are much appreciated :-) Would you have 
 time to review the documentation at 
 https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated by 
 the problem you had.

 Cheers

 On 17/12/2014 14:03, Italo Santos wrote:
 Hello Loic,

 Thanks for you help, I’ve take a look to my crush map and I replace step 
 chooseleaf indep 0 type osd” by step choose indep 0 type osd” and all 
 PGs was created successfully.

 At.

 *Italo Santos*
 http://italosantos.com.br/

 On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:

 Hi,

 The 2147483647 means that CRUSH did not find enough OSD for a given PG. 
 If you check the crush rule associated with the erasure coded pool, you 
 will most probably find why.

 Cheers

 On 16/12/2014 23:32, Italo Santos wrote:
 Hello,

 I'm trying to create an erasure pool following 
 http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but 
 when I try create a pool with a specifc erasure-code-profile 
 (myprofile) the PGs became on incomplete state.

 Anyone can help me?

 Below the profile I created:
 root@ceph0001:~# ceph osd erasure-code-profile get myprofile
 directory=/usr/lib/ceph/erasure-code
 k=6
 m=2
 plugin=jerasure
 technique=reed_sol_van

 The status of cluster:
 root@ceph0001:~# ceph health
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
 unclean

 health detail:
 root@ceph0001:~# ceph health detail
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
 unclean
 pg 2.9 is stuck inactive since forever, current state incomplete, last 
 acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck inactive since forever, current state incomplete, last 
 acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck inactive since forever, current state incomplete, last 
 acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck inactive since forever, current state incomplete, last 
 acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
 pg 2.5 is stuck inactive since forever, current state incomplete, last 
 acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
 pg 2.4 is stuck inactive since forever, current state incomplete, last 
 acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
 pg 2.7 is stuck inactive since forever, current state incomplete, last 
 acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
 pg 2.6 is stuck inactive since forever, current state incomplete, last 
 acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
 pg 2.1 is stuck inactive since forever, current state incomplete, last 
 acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
 pg 2.0 is stuck inactive since forever, current state incomplete, last 
 acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
 pg 2.3 is stuck inactive since forever, current state incomplete, last 
 acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
 pg 2.2 is stuck inactive since forever, current state incomplete, last 
 acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
 pg 2.9 is stuck unclean since forever, current state incomplete, last 
 acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck unclean since forever, current state incomplete, last 
 acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck unclean since forever, current state incomplete, last 
 acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck unclean since forever, current state incomplete, last 
 acting 

Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Italo Santos
Understood.
Thanks for your help, the cluster is healthy now :D

Also, using for example k=6,m=1 and failure domain by host I’ll be able lose 
all OSD on the same host, but if a lose 2 disks on different hosts I can lose 
data right? So, it is possible been a failure domain which allow me to lose an 
OSD or a host?  

Regards.

Italo Santos
http://italosantos.com.br/


On Wednesday, December 17, 2014 at 4:27 PM, Loic Dachary wrote:

  
  
 On 17/12/2014 19:22, Italo Santos wrote:
  Loic,
   
  So, if want have a failure domain by host, I’ll need set up a erasure 
  profile which k+m = total number of hosts I have, right?
  
 Yes, k+m has to be = number of hosts.
  
   
  Regards.
   
  *Italo Santos*
  http://italosantos.com.br/
   
  On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:
   


   On 17/12/2014 18:18, Italo Santos wrote:
Hello,
 
I’ve take a look to this documentation (which help a lot) and if I 
understand right, when I set a profile like:
 
===
ceph osd erasure-code-profile set isilon k=8 m=2 
ruleset-failure-domain=host
===
 
And create a pool following the recommendations on doc, I’ll need 
(100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to 
support create total PGs?

   You will need k+m = 10 host per OSD. If you only have 10 hosts that 
   should be ok and the 800 PGs will use these 10 OSD in various orders. It 
   also means that you will end up having 800 PG per OSD which is a bit too 
   mche. If you have 20 OSDs that will be better : each PG will get 10 OSD 
   out of 20 and each OSD will have 400 PGs. Ideally you want the number of 
   PG per OSD to be in the range (approximately) [20,300].

   Cheers

 
Regards.
 
*Italo Santos*
http://italosantos.com.br/
 
On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:
 
 Hi,
  
 Thanks for the update : good news are much appreciated :-) Would you 
 have time to review the documentation at 
 https://github.com/ceph/ceph/pull/3194/files ? It was partly 
 motivated by the problem you had.
  
 Cheers
  
 On 17/12/2014 14:03, Italo Santos wrote:
  Hello Loic,
   
  Thanks for you help, I’ve take a look to my crush map and I replace 
  step chooseleaf indep 0 type osd” by step choose indep 0 type 
  osd” and all PGs was created successfully.
   
  At.
   
  *Italo Santos*
  http://italosantos.com.br/
   
  On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
   
   Hi,

   The 2147483647 means that CRUSH did not find enough OSD for a 
   given PG. If you check the crush rule associated with the erasure 
   coded pool, you will most probably find why.

   Cheers

   On 16/12/2014 23:32, Italo Santos wrote:
Hello,
 
I'm trying to create an erasure pool following 
http://docs.ceph.com/docs/master/rados/operations/erasure-code/,
 but when I try create a pool with a specifc 
erasure-code-profile (myprofile) the PGs became on incomplete 
state.
 
Anyone can help me?
 
Below the profile I created:
root@ceph0001:~# ceph osd erasure-code-profile get myprofile
directory=/usr/lib/ceph/erasure-code
k=6
m=2
plugin=jerasure
technique=reed_sol_van
 
The status of cluster:
root@ceph0001:~# ceph health
HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs 
stuck unclean
 
health detail:
root@ceph0001:~# ceph health detail
HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs 
stuck unclean
pg 2.9 is stuck inactive since forever, current state 
incomplete, last acting 
[4,10,15,2147483647,3,2147483647,2147483647,2147483647]
pg 2.8 is stuck inactive since forever, current state 
incomplete, last acting 
[0,2147483647,4,2147483647,10,2147483647,15,2147483647]
pg 2.b is stuck inactive since forever, current state 
incomplete, last acting 
[8,3,14,2147483647,5,2147483647,2147483647,2147483647]
pg 2.a is stuck inactive since forever, current state 
incomplete, last acting 
[11,7,2,2147483647,2147483647,2147483647,15,2147483647]
pg 2.5 is stuck inactive since forever, current state 
incomplete, last acting 
[12,8,5,1,2147483647,2147483647,2147483647,2147483647]
pg 2.4 is stuck inactive since forever, current state 
incomplete, last acting 
[5,2147483647,13,1,2147483647,2147483647,8,2147483647]
pg 2.7 is stuck inactive since forever, current state 
incomplete, last acting 
[12,2,10,7,2147483647,2147483647,2147483647,2147483647]
pg 2.6 is stuck inactive since forever, current state 

[ceph-users] Frozen Erasure-coded-pool

2014-12-17 Thread Max Power
I am trying to setup a small VM ceph cluster to excersise before creating a real
cluster. Currently there are two osd's on the same host. I wanted to create an
erasure coded pool with k=1 and m=1 (yes I know it's stupid, but it is a test
case). On top of it there is a cache tier (writeback) and I used the pool to
make a rados blockdevice with it. But as I wanted to format it with ext4 the
system suddenly hangs. At the moment I do not understand why.

I discovered that after the creation of the 'cold-storage' the active primaries
are setup correctly (about one half of the pgs to osd.0 and the other half to
osd.1). But the second osd in the active group is always nonsense (MAXINT, a
placeholder for 'not there'?). To my suprise the state is 'active+clean' - how
can this be, shouldn't it be 'active+degraded'?

These are the commands I used (from my recollection)
:# ceph osd erasure-code-profile get ec_1_1
 directory=/usr/lib/x86_64-linux-gnu/ceph/erasure-code
 k=1
 m=1
 plugin=jerasure
 ruleset-failure-domain=osd
 technique=reed_sol_van
:# ceph osd pool create liverpool 300 300 erasure ec_1_1
:# ceph osd pool create cache 100 100 replicated
:# ceph osd tier add liverpool cache
:# ceph osd tier cache-mode writeback
:# ceph osd tier set-overlay liverpool cache
:# rbd --pool liverpool create --size 1500 testdisk
:# rbd --pool liverpool map testdisk
:# mkfs.ext4 /dev/rbd/liverpool/testdisk

Now the mkfs freezes and I can see this thru ceph -w:
2014-12-17 19:08:56.466846 mon.0 [INF] pgmap v2062: 400 pgs: 400 active+clean;
140 bytes data, 88220 kB used, 2418 MB / 2504 MB avail; 47 B/s rd, 0 op/s
2014-12-17 19:11:20.697190 mon.0 [INF] pgmap v2064: 400 pgs: 307
stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 2504
MB avail
2014-12-17 19:11:20.388468 osd.1 [WRN] 6 slow requests, 6 included below; oldest
blocked for  124.270960 secs
2014-12-17 19:11:20.388556 osd.1 [WRN] slow request 124.270960 seconds old,
received at 2014-12-17 19:09:16.116251: osd_op(client.6155.1:508
rb.0.1807.2ae8944a.0005 [set-alloc-hint object_size 4194304 write_size
4194304,write 4091904~24576] 24.e6ca00e6 ondisk+write e590) v4 currently waiting
for subops from 0
[repeated a few times]
2014-12-17 19:11:21.911696 mon.0 [INF] osdmap e592: 2 osds: 1 up, 2 in
2014-12-17 19:11:22.053272 mon.0 [INF] pgmap v2065: 400 pgs: 307
stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 2504
MB avail
2014-12-17 19:11:24.826008 mon.0 [INF] osd.0 10.0.0.141:6800/7919 boot
2014-12-17 19:11:24.827218 mon.0 [INF] osdmap e593: 2 osds: 2 up, 2 in
2014-12-17 19:11:24.935173 mon.0 [INF] pgmap v2066: 400 pgs: 307
stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 2504
MB avail
2014-12-17 19:11:26.072303 mon.0 [INF] osdmap e594: 2 osds: 2 up, 2 in
2014-12-17 19:11:26.220102 mon.0 [INF] pgmap v2067: 400 pgs: 307
stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 2504
MB avail
2014-12-17 19:11:30.702281 mon.0 [INF] pgmap v2068: 400 pgs: 307
stale+active+clean, 93 active+clean; 16308 kB data, 138 MB used, 2366 MB / 2504
MB avail; 1471 kB/s wr, 7 op/s; 2184 kB/s, 0 objects/s recovering
2014-12-17 19:11:32.050330 mon.0 [INF] pgmap v2069: 400 pgs: 400 active+clean;
33924 kB data, 167 MB used, 2337 MB / 2504 MB avail; 4543 kB/s wr, 46 op/s; 3565
kB/s, 1 objects/s recovering
2014-12-17 19:13:30.569447 mon.0 [INF] pgmap v2070: 400 pgs: 400 active+clean;
33924 kB data, 143 MB used, 2361 MB / 2504 MB avail

How is this explained? What have I done wrong?

Greetings!
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Frozen Erasure-coded-pool

2014-12-17 Thread Loic Dachary
Hi Max,

On 17/12/2014 20:57, Max Power wrote:
 I am trying to setup a small VM ceph cluster to excersise before creating a 
 real
 cluster. Currently there are two osd's on the same host. I wanted to create an
 erasure coded pool with k=1 and m=1 (yes I know it's stupid, but it is a test
 case). 

This is going to fail in an undefined way and it should fail early instead of 
allowing it to proceed. I've created http://tracker.ceph.com/issues/10358, 
thanks for catching this :-) I'm not sure what to make of the rest of what you 
observed, it probably is tainted by the fact that the erasure coded pool cannot 
proceed.

Cheers

 On top of it there is a cache tier (writeback) and I used the pool to
 make a rados blockdevice with it. But as I wanted to format it with ext4 the
 system suddenly hangs. At the moment I do not understand why.
 
 I discovered that after the creation of the 'cold-storage' the active 
 primaries
 are setup correctly (about one half of the pgs to osd.0 and the other half to
 osd.1). But the second osd in the active group is always nonsense (MAXINT, a
 placeholder for 'not there'?). To my suprise the state is 'active+clean' - how
 can this be, shouldn't it be 'active+degraded'?
 
 These are the commands I used (from my recollection)
 :# ceph osd erasure-code-profile get ec_1_1
 directory=/usr/lib/x86_64-linux-gnu/ceph/erasure-code
 k=1
 m=1
 plugin=jerasure
 ruleset-failure-domain=osd
 technique=reed_sol_van
 :# ceph osd pool create liverpool 300 300 erasure ec_1_1
 :# ceph osd pool create cache 100 100 replicated
 :# ceph osd tier add liverpool cache
 :# ceph osd tier cache-mode writeback
 :# ceph osd tier set-overlay liverpool cache
 :# rbd --pool liverpool create --size 1500 testdisk
 :# rbd --pool liverpool map testdisk
 :# mkfs.ext4 /dev/rbd/liverpool/testdisk
 
 Now the mkfs freezes and I can see this thru ceph -w:
 2014-12-17 19:08:56.466846 mon.0 [INF] pgmap v2062: 400 pgs: 400 active+clean;
 140 bytes data, 88220 kB used, 2418 MB / 2504 MB avail; 47 B/s rd, 0 op/s
 2014-12-17 19:11:20.697190 mon.0 [INF] pgmap v2064: 400 pgs: 307
 stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 
 2504
 MB avail
 2014-12-17 19:11:20.388468 osd.1 [WRN] 6 slow requests, 6 included below; 
 oldest
 blocked for  124.270960 secs
 2014-12-17 19:11:20.388556 osd.1 [WRN] slow request 124.270960 seconds old,
 received at 2014-12-17 19:09:16.116251: osd_op(client.6155.1:508
 rb.0.1807.2ae8944a.0005 [set-alloc-hint object_size 4194304 write_size
 4194304,write 4091904~24576] 24.e6ca00e6 ondisk+write e590) v4 currently 
 waiting
 for subops from 0
 [repeated a few times]
 2014-12-17 19:11:21.911696 mon.0 [INF] osdmap e592: 2 osds: 1 up, 2 in
 2014-12-17 19:11:22.053272 mon.0 [INF] pgmap v2065: 400 pgs: 307
 stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 
 2504
 MB avail
 2014-12-17 19:11:24.826008 mon.0 [INF] osd.0 10.0.0.141:6800/7919 boot
 2014-12-17 19:11:24.827218 mon.0 [INF] osdmap e593: 2 osds: 2 up, 2 in
 2014-12-17 19:11:24.935173 mon.0 [INF] pgmap v2066: 400 pgs: 307
 stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 
 2504
 MB avail
 2014-12-17 19:11:26.072303 mon.0 [INF] osdmap e594: 2 osds: 2 up, 2 in
 2014-12-17 19:11:26.220102 mon.0 [INF] pgmap v2067: 400 pgs: 307
 stale+active+clean, 93 active+clean; 140 bytes data, 106 MB used, 2397 MB / 
 2504
 MB avail
 2014-12-17 19:11:30.702281 mon.0 [INF] pgmap v2068: 400 pgs: 307
 stale+active+clean, 93 active+clean; 16308 kB data, 138 MB used, 2366 MB / 
 2504
 MB avail; 1471 kB/s wr, 7 op/s; 2184 kB/s, 0 objects/s recovering
 2014-12-17 19:11:32.050330 mon.0 [INF] pgmap v2069: 400 pgs: 400 active+clean;
 33924 kB data, 167 MB used, 2337 MB / 2504 MB avail; 4543 kB/s wr, 46 op/s; 
 3565
 kB/s, 1 objects/s recovering
 2014-12-17 19:13:30.569447 mon.0 [INF] pgmap v2070: 400 pgs: 400 active+clean;
 33924 kB data, 143 MB used, 2361 MB / 2504 MB avail
 
 How is this explained? What have I done wrong?
 
 Greetings!
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
 

-- 
Loïc Dachary, Artisan Logiciel Libre



signature.asc
Description: OpenPGP digital signature
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Erasure coded PGs incomplete

2014-12-17 Thread Loic Dachary


On 17/12/2014 19:46, Italo Santos wrote: Understood.
 Thanks for your help, the cluster is healthy now :D
 
 Also, using for example k=6,m=1 and failure domain by host I’ll be able lose 
 all OSD on the same host, but if a lose 2 disks on different hosts I can lose 
 data right? So, it is possible been a failure domain which allow me to lose 
 an OSD or a host?

That's actually a good way to put it :-)

 
 Regards.
 
 *Italo Santos*
 http://italosantos.com.br/
 
 On Wednesday, December 17, 2014 at 4:27 PM, Loic Dachary wrote:
 


 On 17/12/2014 19:22, Italo Santos wrote:
 Loic,

 So, if want have a failure domain by host, I’ll need set up a erasure 
 profile which k+m = total number of hosts I have, right?

 Yes, k+m has to be = number of hosts.


 Regards.

 *Italo Santos*
 http://italosantos.com.br/

 On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:



 On 17/12/2014 18:18, Italo Santos wrote:
 Hello,

 I’ve take a look to this documentation (which help a lot) and if I 
 understand right, when I set a profile like:

 ===
 ceph osd erasure-code-profile set isilon k=8 m=2 
 ruleset-failure-domain=host
 ===

 And create a pool following the recommendations on doc, I’ll need 
 (100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to support 
 create total PGs?

 You will need k+m = 10 host per OSD. If you only have 10 hosts that should 
 be ok and the 800 PGs will use these 10 OSD in various orders. It also 
 means that you will end up having 800 PG per OSD which is a bit too mche. 
 If you have 20 OSDs that will be better : each PG will get 10 OSD out of 
 20 and each OSD will have 400 PGs. Ideally you want the number of PG per 
 OSD to be in the range (approximately) [20,300].

 Cheers


 Regards.

 *Italo Santos*
 http://italosantos.com.br/

 On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:

 Hi,

 Thanks for the update : good news are much appreciated :-) Would you 
 have time to review the documentation at 
 https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated 
 by the problem you had.

 Cheers

 On 17/12/2014 14:03, Italo Santos wrote:
 Hello Loic,

 Thanks for you help, I’ve take a look to my crush map and I replace 
 step chooseleaf indep 0 type osd” by step choose indep 0 type osd” 
 and all PGs was created successfully.

 At.

 *Italo Santos*
 http://italosantos.com.br/

 On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:

 Hi,

 The 2147483647 means that CRUSH did not find enough OSD for a given 
 PG. If you check the crush rule associated with the erasure coded 
 pool, you will most probably find why.

 Cheers

 On 16/12/2014 23:32, Italo Santos wrote:
 Hello,

 I'm trying to create an erasure pool following 
 http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but 
 when I try create a pool with a specifc erasure-code-profile 
 (myprofile) the PGs became on incomplete state.

 Anyone can help me?

 Below the profile I created:
 root@ceph0001:~# ceph osd erasure-code-profile get myprofile
 directory=/usr/lib/ceph/erasure-code
 k=6
 m=2
 plugin=jerasure
 technique=reed_sol_van

 The status of cluster:
 root@ceph0001:~# ceph health
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
 unclean

 health detail:
 root@ceph0001:~# ceph health detail
 HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
 unclean
 pg 2.9 is stuck inactive since forever, current state incomplete, 
 last acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
 pg 2.8 is stuck inactive since forever, current state incomplete, 
 last acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
 pg 2.b is stuck inactive since forever, current state incomplete, 
 last acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
 pg 2.a is stuck inactive since forever, current state incomplete, 
 last acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
 pg 2.5 is stuck inactive since forever, current state incomplete, 
 last acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
 pg 2.4 is stuck inactive since forever, current state incomplete, 
 last acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
 pg 2.7 is stuck inactive since forever, current state incomplete, 
 last acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
 pg 2.6 is stuck inactive since forever, current state incomplete, 
 last acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
 pg 2.1 is stuck inactive since forever, current state incomplete, 
 last acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
 pg 2.0 is stuck inactive since forever, current state incomplete, 
 last acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
 pg 2.3 is stuck inactive since forever, current state incomplete, 
 last acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
 pg 2.2 is stuck inactive since forever, current state incomplete, 
 last acting 

[ceph-users] Double-mounting of RBD

2014-12-17 Thread McNamara, Bradley
I have a somewhat interesting scenario.  I have an RBD of 17TB formatted using 
XFS.  I would like it accessible from two different hosts, one mapped/mounted 
read-only, and one mapped/mounted as read-write.  Both are shared using Samba 
4.x.  One Samba server gives read-only access to the world for the data.  The 
other gives read-write access to a very limited set of users who occasionally 
need to add data.

However, when testing this, when changes are made to the read-write Samba 
server the changes don't seem to be seen by the read-only Samba server.  Is 
there some file system caching going on that will eventually be flushed?

Am I living dangerously doing what I have set up?  I thought I would avoid 
most/all potential file system corruption by making sure there is only one 
read-write access method.  Thanks for any answers.

Brad
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Double-mounting of RBD

2014-12-17 Thread Gregory Farnum
On Wed, Dec 17, 2014 at 2:31 PM, McNamara, Bradley
bradley.mcnam...@seattle.gov wrote:
 I have a somewhat interesting scenario.  I have an RBD of 17TB formatted
 using XFS.  I would like it accessible from two different hosts, one
 mapped/mounted read-only, and one mapped/mounted as read-write.  Both are
 shared using Samba 4.x.  One Samba server gives read-only access to the
 world for the data.  The other gives read-write access to a very limited set
 of users who occasionally need to add data.


 However, when testing this, when changes are made to the read-write Samba
 server the changes don’t seem to be seen by the read-only Samba server.  Is
 there some file system caching going on that will eventually be flushed?



 Am I living dangerously doing what I have set up?  I thought I would avoid
 most/all potential file system corruption by making sure there is only one
 read-write access method.  Thanks for any answers.

Well, you'll avoid corruption by only having one writer, but the other
reader is still caching data in-memory that will prevent it from
seeing the writes on the disk.
Plus I have no idea if mounting xfs read-only actually prevents it
from making any writes to the disk; I think some FSes will do stuff
like defragment internal data structures in that mode, maybe?
-Greg
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Double-mounting of RBD

2014-12-17 Thread Josh Durgin

On 12/17/2014 03:49 PM, Gregory Farnum wrote:

On Wed, Dec 17, 2014 at 2:31 PM, McNamara, Bradley
bradley.mcnam...@seattle.gov wrote:

I have a somewhat interesting scenario.  I have an RBD of 17TB formatted
using XFS.  I would like it accessible from two different hosts, one
mapped/mounted read-only, and one mapped/mounted as read-write.  Both are
shared using Samba 4.x.  One Samba server gives read-only access to the
world for the data.  The other gives read-write access to a very limited set
of users who occasionally need to add data.


However, when testing this, when changes are made to the read-write Samba
server the changes don’t seem to be seen by the read-only Samba server.  Is
there some file system caching going on that will eventually be flushed?



Am I living dangerously doing what I have set up?  I thought I would avoid
most/all potential file system corruption by making sure there is only one
read-write access method.  Thanks for any answers.


Well, you'll avoid corruption by only having one writer, but the other
reader is still caching data in-memory that will prevent it from
seeing the writes on the disk.
Plus I have no idea if mounting xfs read-only actually prevents it
from making any writes to the disk; I think some FSes will do stuff
like defragment internal data structures in that mode, maybe?
-Greg


FSes mounted read-only still do tend to do things like journal replay,
but since the block device is mapped read-only that won't be a problem
in this case.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] File System stripping data

2014-12-17 Thread Kevin Shiah
Hi John,

I am using 0.56.1. Could it be because data striping is not supported in
this version?

Kevin

On Wed Dec 17 2014 at 4:00:15 AM PST Wido den Hollander w...@42on.com
wrote:

 On 12/17/2014 12:35 PM, John Spray wrote:
  On Wed, Dec 17, 2014 at 10:25 AM, Wido den Hollander w...@42on.com
 wrote:
  I just tried something similar on Giant (0.87) and I saw this in the
 logs:
 
  parse_layout_vxattr name layout.pool value 'cephfs_svo'
   invalid data pool 3
  reply request -22
 
  I resolves the pool to a ID, but then it's unable to set it?
 
  Was the 'cephfs_svo' pool already added as a data pool with ceph mds
  add_data_pool?
 

 Ah, indeed. Working fine right now. Same goes for any other layout
 settings.

  There are paths where if a pool was added very recently, MDSs/clients
  might not know about the pool yet and can generate errors like this.
 
  John
 


 --
 Wido den Hollander
 42on B.V.
 Ceph trainer and consultant

 Phone: +31 (0)20 700 9902
 Skype: contact42on
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Help with SSDs

2014-12-17 Thread Christian Balzer

Hello,

On Tue, 16 Dec 2014 08:58:23 -0700 Bryson McCutcheon wrote:

 Hey there,
 
 Is there a good work around if our SSDs are not handling D_SYNC very
 well? We invested a ton of money into Samsung 840 EVOS and they are not
 playing well with D_SYNC.  Would really appreciate the help!
 
Baring any reply from the developers, I'm pretty sure the answer is no, as
D_SYNC is a vital part of the journaling process. 

You _may_ get away with a hardware cache RAID controller (some
models/makers will use that cache in JBOD mode, too).

And while I very much can sympathize with you (I have dozens of new Toshiba
DTxx drives sitting on a pile that perform quite nicely for a week before
they go into half to 10% speed mode for at least 10 hours) and don't want
to sound snarky, penultimately you get what you pay for. 

Journals on consumer SSDs are a bad idea. Aside form the performance
issues you're experiencing, Samsung doesn't really give a TBW value for
them, the estimation based on their claims is about 100TBW for the 120GB
model. While not bad, it's something I'd be weary about with regards to
long term usage.

http://ssdendurancetest.com/ssd-endurance-test-report/Samsung-840-EVO-120

The only reliable game in town at this point in time are Intel DC S3700
models, the 200GB model for example has a TBW of 1.8PB and will keep
its speed w/o the need for TRIM or massive underprovisioning.  

Christian
-- 
Christian BalzerNetwork/Systems Engineer
ch...@gol.com   Global OnLine Japan/Fusion Communications
http://www.gol.com/
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Double-mounting of RBD

2014-12-17 Thread Jake Young
On Wednesday, December 17, 2014, Josh Durgin josh.dur...@inktank.com
wrote:

 On 12/17/2014 03:49 PM, Gregory Farnum wrote:

 On Wed, Dec 17, 2014 at 2:31 PM, McNamara, Bradley
 bradley.mcnam...@seattle.gov wrote:

 I have a somewhat interesting scenario.  I have an RBD of 17TB formatted
 using XFS.  I would like it accessible from two different hosts, one
 mapped/mounted read-only, and one mapped/mounted as read-write.  Both are
 shared using Samba 4.x.  One Samba server gives read-only access to the
 world for the data.  The other gives read-write access to a very limited
 set
 of users who occasionally need to add data.


 However, when testing this, when changes are made to the read-write Samba
 server the changes don’t seem to be seen by the read-only Samba server.
 Is
 there some file system caching going on that will eventually be flushed?



 Am I living dangerously doing what I have set up?  I thought I would
 avoid
 most/all potential file system corruption by making sure there is only
 one
 read-write access method.  Thanks for any answers.


 Well, you'll avoid corruption by only having one writer, but the other
 reader is still caching data in-memory that will prevent it from
 seeing the writes on the disk.
 Plus I have no idea if mounting xfs read-only actually prevents it
 from making any writes to the disk; I think some FSes will do stuff
 like defragment internal data structures in that mode, maybe?
 -Greg


 FSes mounted read-only still do tend to do things like journal replay,
 but since the block device is mapped read-only that won't be a problem
 in this case.
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Someone commented that the OS with the readonly mount will still do
something potentially damaging to the filesystem at mount time. Something
along the lines of replaying the xfs journal and the read write OS being
unaware of it.

Dig through the ceph mailing list archives.

Jake
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Help with SSDs

2014-12-17 Thread Mikaël Cluseau

On 12/17/2014 02:58 AM, Bryson McCutcheon wrote:
Is there a good work around if our SSDs are not handling D_SYNC very 
well? We invested a ton of money into Samsung 840 EVOS and they are 
not playing well with D_SYNC.  Would really appreciate the help!


Just in case it's linked with the recent performance issue: are they 
upgraded/refreshed with the latest firmware?

(http://www.samsung.com/au/consumer/pc-peripherals/solid-state-drive/ssd-840-evo/MZ-7TE1T0BW)

I have EVOs too, what to you mean by not playing well with D_SYNC? Is 
there something I can test on my side to compare results with you, as I 
have mine flashed?

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Reproducable Data Corruption with cephfs kernel driver

2014-12-17 Thread Lindsay Mathieson
I'be been experimenting with CephFS for funning KVM images (proxmox).

cephfs fuse version - 0.87

cephfs kernel module - kernel version 3.10


Part of my testing involves running a Windows 7 VM up and running
CrystalDiskMark to check the I/O in the VM. Its surprisingly good with
both the fuse and the kernel driver, seq reads  writes are actually
faster than the underlying disk, so I presume the FS is aggressively
caching.

With the fuse driver I have no problems.

With the kernel driver, the benchmark runs fine, but when I reboot the
VM the drive is corrupted and unreadable, every time. Rolling back to
a snapshot fixes the disk. This does not happen unless I run the
benchmark, which I presume is writing a lot of data.

No problems with the same test for Ceph rbd, or NFS.


-- 
Lindsay
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Help with SSDs

2014-12-17 Thread Udo Lembke
Hi Mikaël,


 I have EVOs too, what to you mean by not playing well with D_SYNC?
 Is there something I can test on my side to compare results with you,
 as I have mine flashed?
http://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/
described how test the ssd-performance for an journal ssd (your ssd will
be overwritten!!).

Udo
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Help with SSDs

2014-12-17 Thread Alexandre DERUMIER
what to you mean by not playing well with D_SYNC?

Hi, check this blog:
http://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/


- Mail original -
De: Mikaël Cluseau mclus...@isi.nc
À: Bryson McCutcheon brysonmccutch...@gmail.com, ceph-users 
ceph-users@lists.ceph.com
Envoyé: Jeudi 18 Décembre 2014 03:16:07
Objet: Re: [ceph-users] Help with SSDs

On 12/17/2014 02:58 AM, Bryson McCutcheon wrote: 
 Is there a good work around if our SSDs are not handling D_SYNC very 
 well? We invested a ton of money into Samsung 840 EVOS and they are 
 not playing well with D_SYNC. Would really appreciate the help! 

Just in case it's linked with the recent performance issue: are they 
upgraded/refreshed with the latest firmware? 
(http://www.samsung.com/au/consumer/pc-peripherals/solid-state-drive/ssd-840-evo/MZ-7TE1T0BW)
 

I have EVOs too, what to you mean by not playing well with D_SYNC? Is 
there something I can test on my side to compare results with you, as I 
have mine flashed? 
___ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Help with SSDs

2014-12-17 Thread Mark Kirkwood
Looking at the blog, I notice he disabled the write cache before the 
tests: doing this on my m550 resulted in *improved* dsync results (300 
IOPS - 700 IOPS) still not great obviously, but ... interesting.


So do experiment with the settings to see if you can get the 840's 
working better for you.


While you can't do much about the endurance lifetime being a bit low, 
you could possibly improve performance using a journal *file* that is 
located on the 840's (you'll need to symlink it - disclaimer - have not 
tried this myself, but will experiment if you are interested). Slightly 
different open() options are used in this case and these cheaper 
consumer SSD seem to work better with them.


Regards

Mark

On 18/12/14 18:08, Alexandre DERUMIER wrote:

what to you mean by not playing well with D_SYNC?


Hi, check this blog:
http://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/



___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] 'rbd list' stuck

2014-12-17 Thread yang . bin18
The cluster state must be wrong,but how to recovery?

root@node3 ceph-cluster]# ceph -w
cluster 1365f2dd-b86c-436c-a64f-3318a937f3c2
 health HEALTH_WARN 64 pgs incomplete; 64 pgs stale; 64 pgs stuck 
inactive; 64 pgs stuck stale; 64 pgs stuck unclean; 8 requests are blocked 
 32 sec
 monmap e1: 1 mons at {node3=10.74.120.79:6789/0}, election epoch 1, 
quorum 0 node3
 osdmap e38: 3 osds: 3 up, 3 in
  pgmap v80: 64 pgs, 1 pools, 0 bytes data, 0 objects
100 MB used, 164 GB / 164 GB avail
  64 stale+incomplete

2014-12-18 09:51:44.156184 mon.0 [INF] pgmap v80: 64 pgs: 64 
stale+incomplete; 0 bytes data, 100 MB used, 164 GB / 164 GB avail




Hi

Why command 'rbd list'executed on monitor stuck,any prompt should be 
appreciated!

Backtree: 


[810bfdee] futex_wait_queue_me+0xde/0x140
[810c0969] futex_wait+0x179/0x280
[810c297e] do_futex+0xfe/0x5e0
[810c2ee0] SyS_futex+0x80/0x180
[815f2119] system_call_fastpath+0x16/0x1b
[] 0x

Best Regards!

 YangBin


ZTE Information Security Notice: The information contained in this mail (and 
any attachment transmitted herewith) is privileged and confidential and is 
intended for the exclusive use of the addressee(s).  If you are not an intended 
recipient, any disclosure, reproduction, distribution or other dissemination or 
use of the information contained is strictly prohibited.  If you have received 
this mail in error, please delete it and notify us immediately.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com