[konetzed@ovz2 ~]$ sudo pstorage -c test_cluster stat
connected to MDS#3
Cluster 'test_cluster': healthy
Space: [OK] allocatable 28GB of 35GB, free 31GB of 35GB
MDS nodes: 3 of 3, epoch uptime: 10h 25m
CS nodes: 3 of 3 (3 avail, 0 inactive, 0 offline)
License: [Error] License not loaded, capacity limited to 100Gb
Replication: 1 norm, 1 limit
Chunks: [OK] 1 (100%) healthy, 0 (0%) standby, 0 (0%) degraded, 0 (0%)
urgent,
0 (0%) blocked, 0 (0%) pending, 0 (0%) offline, 0 (0%)
replicating,
0 (0%) overcommitted, 0 (0%) deleting, 0 (0%) void
FS: 10KB in 2 files, 2 inodes, 1 file maps, 1 chunks, 1 chunk replicas
IO: read 0B/s ( 0ops/s), write 0B/s ( 0ops/s)
IO total: read 0B ( 0ops), write 0B ( 0ops)
Repl IO: read 0B/s, write: 0B/s
Sync rate: 0ops/s, datasync rate: 0ops/s
MDSID STATUS %CTIME COMMITS %CPU MEM UPTIME HOST
1 avail 3.1% 1/s 0.1% 14m 9h 58m ovz1.home.int:2510
2 avail 2.5% 0/s 0.0% 14m 9h 14m ovz2.home.int:2510
M 3 avail 3.0% 1/s 0.3% 15m 10h 25m ovz3.home.int:2510
CSID STATUS SPACE FREE REPLICAS IOWAIT IOLAT(ms) QDEPTH HOST
1025 active 11GB 10GB 0 0% 0/0 0.0
ovz1.home.int
1026 active 11GB 10GB 0 0% 0/0 0.0
ovz2.home.int
1027 active 11GB 10GB 1 0% 0/0 0.0
ovz3.home.int
CLID LEASES READ WRITE RD_OPS WR_OPS FSYNCS
IOLAT(ms) HOST
2060 0/0 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0
ovz3.home.int
2065 0/1 0B/s 0B/s 0ops/s 0ops/s 0ops/s 0/0
ovz1.home.int
I do have skype but I have meetings all day for work and cant be on a
computer after. I may have time tomorrow if that would work. I am in
the central time zone.
Edward
On 01/29/2014 03:14 AM, Kirill Korotaev wrote:
Edward,
can you send me in private email output of:
# pstorage -c <cluster> stat
output?
Do you have a skype?
Thanks,
Kirill
On 29 Jan 2014, at 10:26, Edward Konetzko <konet...@gmail.com
<mailto:konet...@gmail.com>> wrote:
On 01/28/2014 09:51 AM, Kir Kolyshkin wrote:
On 28 January 2014 02:55, Kirill Korotaev <d...@parallels.com
<mailto:d...@parallels.com>> wrote:
>> On 25 Jan 2014, at 07:38, Rene C. ope...@dokbua.com
<mailto:ope...@dokbua.com> wrote:
>>
>
> Hi,
>
> I read the website about the cloud storage and I found some
words, which seems familiar for me.
>
> May I ask, which filesystem do you use to be able to regularly
scrub and self-heal the filesystem?
>
> Personaly I use zfsonlinux in production for a long time now
and I am very satisfied with it, and based on your description,
it seems you should use something like that and something on top
of the native filesystem to get a cloud storage.
>
> Or you use a ceph or alike "filesystem", which has similar
capabilities with cloud features.
It’s more like a ceph. Data is stored in a distributed way, so
unlike to zfs you have access to the data even in case of node
failure (crash, CPU/memory fault etc.) and access is available
from ANY cluster node.
As such we store the data and maintain checksums on every node
and can do periodic scrubbing of the data.
Just to clarify -- this is Parallels own distributed/cloud
filesystem, not CEPH or GlusterFS,
but similar to. For more info, check the links at
https://openvz.org/Parallels_Cloud_Storage#External_links
_______________________________________________
Users mailing list
Users@openvz.org
https://lists.openvz.org/mailman/listinfo/users
Setup a cluster using Centos 6.5 64bit, fresh install in KVM
instances. I wanted to test functionality not actual speed.
All software was latest as of last night and I followed the quick how
to here https://openvz.org/Parallels_Cloud_Storage
Everything works great until I try to create an instance using the
command "vzctl create 101 --layout ploop --ostemplate centos-6-x86_64
--private /pcs/containers/101" from the docs.
About one mb of data is written to disk and then it just hangs. The
following is output from dmesg
[ 360.414242] INFO: task vzctl:1646 blocked for more than 120 seconds.
[ 360.414770] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs"
disables this message.
[ 360.415406] vzctl D ffff88007e444500 0 1646 1611
0 0x00000084
[ 360.415418] ffff88007ea59a68 0000000000000086 ffff8800ffffffff
000006b62934b8c0
[ 360.415428] 0000000000000000 ffff88007e9f2ad0 0000000000005eaa
ffffffffad17694d
[ 360.415437] 000000000ad7ef74 ffffffff81a97b40 ffff88007e444ac8
000000000001eb80
[ 360.415452] Call Trace:
[ 360.415492] [<ffffffff81517353>] io_schedule+0x73/0xc0
[ 360.415516] [<ffffffff811f39b3>] wait_on_sync_kiocb+0x53/0x80
[ 360.415537] [<ffffffffa04dbf47>] fuse_direct_IO+0x167/0x230 [fuse]
[ 360.415558] [<ffffffff8112e948>] mapping_direct_IO+0x48/0x70
[ 360.415567] [<ffffffff811301a6>]
generic_file_direct_write_iter+0xf6/0x170
[ 360.415576] [<ffffffff81130c8e>]
__generic_file_write_iter+0x32e/0x420
[ 360.415585] [<ffffffff81130e05>] __generic_file_aio_write+0x85/0xa0
[ 360.415594] [<ffffffff81130ea8>] generic_file_aio_write+0x88/0x100
[ 360.415605] [<ffffffffa04da085>] fuse_file_aio_write+0x185/0x430
[fuse]
[ 360.415623] [<ffffffff811a530a>] do_sync_write+0xfa/0x140
[ 360.415641] [<ffffffff8109d930>] ? autoremove_wake_function+0x0/0x40
[ 360.415655] [<ffffffff812902da>] ? strncpy_from_user+0x4a/0x90
[ 360.415664] [<ffffffff811a55e8>] vfs_write+0xb8/0x1a0
[ 360.415671] [<ffffffff811a5ee1>] sys_write+0x51/0x90
[ 360.415681] [<ffffffff8100b102>] system_call_fastpath+0x16/0x1b
Even just trying to create a 10k file with dd causes a task to hang.
"dd if=/dev/zero of=/pcs/test.junk bs=1k count=10"
Any ideas? Anymore info you would like for debugging.
_______________________________________________
Users mailing list
Users@openvz.org <mailto:Users@openvz.org>
https://lists.openvz.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users@openvz.org
https://lists.openvz.org/mailman/listinfo/users