Added 10G to each vm and rebooted, got the same issue. Included the same output you asked for last time to see if there was something obvious.

[root@ovz3 ~]# pstorage -c test_cluster stat
connected to MDS#2
Cluster 'test_cluster': healthy
Space: [OK] allocatable 56GB of 63GB, free 60GB of 63GB
MDS nodes: 3 of 3, epoch uptime: 2h 20m
CS nodes:  3 of 3 (3 avail, 0 inactive, 0 offline)
License: [Error] License not loaded, capacity limited to 100Gb
Replication:  1 norm,  1 limit
Chunks: [OK] 1 (100%) healthy, 0 (0%) standby, 0 (0%) degraded, 0 (0%) urgent, 0 (0%) blocked, 0 (0%) pending, 0 (0%) offline, 0 (0%) replicating,
             0 (0%) overcommitted,  0 (0%) deleting,  0 (0%) void
FS:  1MB in 4 files, 4 inodes,  1 file maps,  1 chunks,  1 chunk replicas
IO:       read     0B/s (  0ops/s), write     0B/s (  0ops/s)
IO total: read       0B (    0ops), write       0B (    0ops)
Repl IO:  read     0B/s, write:     0B/s
Sync rate:   0ops/s, datasync rate:   0ops/s

MDSID STATUS   %CTIME   COMMITS   %CPU    MEM   UPTIME HOST
    1 avail      2.0%       0/s   0.0%    18m   2h 20m ovz1.home.int:2510
M   2 avail      2.4%       0/s   0.1%    18m   2h 20m ovz2.home.int:2510
    3 avail      3.8%       1/s   0.0%    18m   2h 20m ovz3.home.int:2510

 CSID STATUS      SPACE   FREE REPLICAS IOWAIT IOLAT(ms) QDEPTH HOST
1025 active 21GB 19GB 0 0% 0/0 0.0 ovz1.home.int 1026 active 21GB 19GB 0 0% 0/0 0.0 ovz2.home.int 1027 active 21GB 20GB 1 0% 0/0 0.0 ovz3.home.int

CLID LEASES READ WRITE RD_OPS WR_OPS FSYNCS IOLAT(ms) HOST 2089 0/1 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0 ovz1.home.int 2090 0/0 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0 ovz2.home.int 2091 0/0 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0 ovz3.home.int






On 01/29/2014 01:45 PM, Kirill Korotaev wrote:
Edward, got it - there is a small threshold (10GB) on minimum free space on CS’es (reserved for different cases include recovery),
you have ~10GB per CS so you hit this threshold immediately.

Most likely you run from inside VMs, right? Just increase disk space available to CS then.


On 29 Jan 2014, at 21:04, Edward Konetzko <konet...@gmail.com <mailto:konet...@gmail.com>> wrote:

[konetzed@ovz2 ~]$ sudo pstorage -c test_cluster stat
connected to MDS#3
Cluster 'test_cluster': healthy
Space: [OK] allocatable 28GB of 35GB, free 31GB of 35GB
MDS nodes: 3 of 3, epoch uptime: 10h 25m
CS nodes:  3 of 3 (3 avail, 0 inactive, 0 offline)
License: [Error] License not loaded, capacity limited to 100Gb
Replication:  1 norm,  1 limit
Chunks: [OK] 1 (100%) healthy, 0 (0%) standby, 0 (0%) degraded, 0 (0%) urgent, 0 (0%) blocked, 0 (0%) pending, 0 (0%) offline, 0 (0%) replicating,
             0 (0%) overcommitted,  0 (0%) deleting, 0 (0%) void
FS: 10KB in 2 files, 2 inodes, 1 file maps, 1 chunks, 1 chunk replicas
IO:       read     0B/s (  0ops/s), write     0B/s ( 0ops/s)
IO total: read       0B (    0ops), write       0B (    0ops)
Repl IO:  read     0B/s, write:     0B/s
Sync rate:   0ops/s, datasync rate:   0ops/s

MDSID STATUS   %CTIME   COMMITS   %CPU    MEM   UPTIME HOST
1 avail 3.1% 1/s 0.1% 14m 9h 58m ovz1.home.int:2510 2 avail 2.5% 0/s 0.0% 14m 9h 14m ovz2.home.int:2510 M 3 avail 3.0% 1/s 0.3% 15m 10h 25m ovz3.home.int:2510

 CSID STATUS      SPACE   FREE REPLICAS IOWAIT IOLAT(ms) QDEPTH HOST
 1025 active       11GB   10GB        0     0% 0/0    0.0 ovz1.home.int
 1026 active       11GB   10GB        0     0% 0/0    0.0 ovz2.home.int
 1027 active       11GB   10GB        1     0% 0/0    0.0 ovz3.home.int

CLID LEASES READ WRITE RD_OPS WR_OPS FSYNCS IOLAT(ms) HOST 2060 0/0 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0 ovz3.home.int 2065 0/1 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0 ovz1.home.int

I do have skype but I have meetings all day for work and cant be on a computer after. I may have time tomorrow if that would work. I am in the central time zone.

Edward


On 01/29/2014 03:14 AM, Kirill Korotaev wrote:
Edward,

can you send me in private email output of:
# pstorage -c <cluster> stat
output?

Do you have a skype?

Thanks,
Kirill



On 29 Jan 2014, at 10:26, Edward Konetzko <konet...@gmail.com <mailto:konet...@gmail.com>> wrote:

On 01/28/2014 09:51 AM, Kir Kolyshkin wrote:
On 28 January 2014 02:55, Kirill Korotaev <d...@parallels.com <mailto:d...@parallels.com>> wrote:

    >> On 25 Jan 2014, at 07:38, Rene C. ope...@dokbua.com
    <mailto:ope...@dokbua.com> wrote:
    >>
    >
    > Hi,
    >
    > I read the website about the cloud storage and I found some
    words, which seems familiar for me.
    >
    > May I ask, which filesystem do you use to be able to
    regularly scrub and self-heal the filesystem?
    >
    > Personaly I use zfsonlinux in production for a long time now
    and I am very satisfied with it, and based on your
    description, it seems you should use something like that and
    something on top of the native filesystem to get a cloud storage.
    >
    > Or you use a ceph or alike "filesystem", which has similar
    capabilities with cloud features.

    It’s more like a ceph. Data is stored in a distributed way, so
    unlike to zfs you have access to the data even in case of node
    failure (crash, CPU/memory fault etc.) and access is available
    from ANY cluster node.
    As such we store the data and maintain checksums on every node
    and can do periodic scrubbing of the data.


Just to clarify -- this is Parallels own distributed/cloud filesystem, not CEPH or GlusterFS, but similar to. For more info, check the links at https://openvz.org/Parallels_Cloud_Storage#External_links



_______________________________________________
Users mailing list
Users@openvz.org
https://lists.openvz.org/mailman/listinfo/users
Setup a cluster using Centos 6.5 64bit, fresh install in KVM instances. I wanted to test functionality not actual speed.

All software was latest as of last night and I followed the quick how to here https://openvz.org/Parallels_Cloud_Storage

Everything works great until I try to create an instance using the command "vzctl create 101 --layout ploop --ostemplate centos-6-x86_64 --private /pcs/containers/101" from the docs.

About one mb of data is written to disk and then it just hangs. The following is output from dmesg

[  360.414242] INFO: task vzctl:1646 blocked for more than 120 seconds.
[ 360.414770] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [ 360.415406] vzctl D ffff88007e444500 0 1646 1611 0 0x00000084 [ 360.415418] ffff88007ea59a68 0000000000000086 ffff8800ffffffff 000006b62934b8c0 [ 360.415428] 0000000000000000 ffff88007e9f2ad0 0000000000005eaa ffffffffad17694d [ 360.415437] 000000000ad7ef74 ffffffff81a97b40 ffff88007e444ac8 000000000001eb80
[  360.415452] Call Trace:
[  360.415492]  [<ffffffff81517353>] io_schedule+0x73/0xc0
[  360.415516]  [<ffffffff811f39b3>] wait_on_sync_kiocb+0x53/0x80
[  360.415537]  [<ffffffffa04dbf47>] fuse_direct_IO+0x167/0x230 [fuse]
[  360.415558]  [<ffffffff8112e948>] mapping_direct_IO+0x48/0x70
[ 360.415567] [<ffffffff811301a6>] generic_file_direct_write_iter+0xf6/0x170 [ 360.415576] [<ffffffff81130c8e>] __generic_file_write_iter+0x32e/0x420
[  360.415585]  [<ffffffff81130e05>] __generic_file_aio_write+0x85/0xa0
[  360.415594]  [<ffffffff81130ea8>] generic_file_aio_write+0x88/0x100
[ 360.415605] [<ffffffffa04da085>] fuse_file_aio_write+0x185/0x430 [fuse]
[  360.415623]  [<ffffffff811a530a>] do_sync_write+0xfa/0x140
[ 360.415641] [<ffffffff8109d930>] ? autoremove_wake_function+0x0/0x40
[  360.415655]  [<ffffffff812902da>] ? strncpy_from_user+0x4a/0x90
[  360.415664]  [<ffffffff811a55e8>] vfs_write+0xb8/0x1a0
[  360.415671]  [<ffffffff811a5ee1>] sys_write+0x51/0x90
[  360.415681]  [<ffffffff8100b102>] system_call_fastpath+0x16/0x1b

Even just trying to create a 10k file with dd causes a task to hang. "dd if=/dev/zero of=/pcs/test.junk bs=1k count=10"


Any ideas? Anymore info you would like for debugging.
_______________________________________________
Users mailing list
Users@openvz.org <mailto:Users@openvz.org>
https://lists.openvz.org/mailman/listinfo/users




_______________________________________________
Users mailing list
Users@openvz.org
https://lists.openvz.org/mailman/listinfo/users

Reply via email to