[Users] Cloud Storage for OpenVZ Containers

Edward Konetzko konetzed at gmail.com
Wed Jan 29 09:04:57 PST 2014


[konetzed at ovz2 ~]$ sudo pstorage -c test_cluster stat
connected to MDS#3
Cluster 'test_cluster': healthy
Space: [OK] allocatable 28GB of 35GB, free 31GB of 35GB
MDS nodes: 3 of 3, epoch uptime: 10h 25m
CS nodes:  3 of 3 (3 avail, 0 inactive, 0 offline)
License: [Error] License not loaded, capacity limited to 100Gb
Replication:  1 norm,  1 limit
Chunks: [OK] 1 (100%) healthy,  0 (0%) standby,  0 (0%) degraded, 0 (0%) 
urgent,
              0 (0%) blocked,  0 (0%) pending,  0 (0%) offline,  0 (0%) 
replicating,
              0 (0%) overcommitted,  0 (0%) deleting,  0 (0%) void
FS:  10KB in 2 files, 2 inodes,  1 file maps,  1 chunks,  1 chunk replicas
IO:       read     0B/s (  0ops/s), write     0B/s (  0ops/s)
IO total: read       0B (    0ops), write       0B (    0ops)
Repl IO:  read     0B/s, write:     0B/s
Sync rate:   0ops/s, datasync rate:   0ops/s

MDSID STATUS   %CTIME   COMMITS   %CPU    MEM   UPTIME HOST
     1 avail      3.1%       1/s   0.1%    14m   9h 58m ovz1.home.int:2510
     2 avail      2.5%       0/s   0.0%    14m   9h 14m ovz2.home.int:2510
M   3 avail      3.0%       1/s   0.3%    15m  10h 25m ovz3.home.int:2510

  CSID STATUS      SPACE   FREE REPLICAS IOWAIT IOLAT(ms) QDEPTH HOST
  1025 active       11GB   10GB        0     0%       0/0    0.0 
ovz1.home.int
  1026 active       11GB   10GB        0     0%       0/0    0.0 
ovz2.home.int
  1027 active       11GB   10GB        1     0%       0/0    0.0 
ovz3.home.int

  CLID   LEASES     READ    WRITE     RD_OPS     WR_OPS     FSYNCS 
IOLAT(ms) HOST
  2060      0/0     0B/s     0B/s     0ops/s     0ops/s 0ops/s       0/0 
ovz3.home.int
  2065      0/1     0B/s     0B/s     0ops/s     0ops/s 0ops/s       0/0 
ovz1.home.int

I do have skype but I have meetings all day for work and cant be on a 
computer after.  I may have time tomorrow if that would work.  I am in 
the central time zone.

Edward


On 01/29/2014 03:14 AM, Kirill Korotaev wrote:
> Edward,
>
> can you send me in private email output of:
> # pstorage -c <cluster> stat
> output?
>
> Do you have a skype?
>
> Thanks,
> Kirill
>
>
>
> On 29 Jan 2014, at 10:26, Edward Konetzko <konetzed at gmail.com 
> <mailto:konetzed at gmail.com>> wrote:
>
>> On 01/28/2014 09:51 AM, Kir Kolyshkin wrote:
>>> On 28 January 2014 02:55, Kirill Korotaev <dev at parallels.com 
>>> <mailto:dev at parallels.com>> wrote:
>>>
>>>     >> On 25 Jan 2014, at 07:38, Rene C. openvz at dokbua.com
>>>     <mailto:openvz at dokbua.com> wrote:
>>>     >>
>>>     >
>>>     > Hi,
>>>     >
>>>     > I read the website about the cloud storage and I found some
>>>     words, which seems familiar for me.
>>>     >
>>>     > May I ask, which filesystem do you use to be able to regularly
>>>     scrub and self-heal the filesystem?
>>>     >
>>>     > Personaly I use zfsonlinux in production for a long time now
>>>     and I am very satisfied with it, and based on your description,
>>>     it seems you should use something like that and something on top
>>>     of the native filesystem to get a cloud storage.
>>>     >
>>>     > Or you use a ceph or alike "filesystem", which has similar
>>>     capabilities with cloud features.
>>>
>>>     It’s more like a ceph. Data is stored in a distributed way, so
>>>     unlike to zfs you have access to the data even in case of node
>>>     failure (crash, CPU/memory fault etc.) and access is available
>>>     from ANY cluster node.
>>>     As such we store the data and maintain checksums on every node
>>>     and can do periodic scrubbing of the data.
>>>
>>>
>>> Just to clarify -- this is Parallels own distributed/cloud 
>>> filesystem, not CEPH or GlusterFS,
>>> but similar to. For more info, check the links at 
>>> https://openvz.org/Parallels_Cloud_Storage#External_links
>>>
>>>
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users at openvz.org
>>> https://lists.openvz.org/mailman/listinfo/users
>> Setup a cluster using Centos 6.5 64bit, fresh install in KVM 
>> instances.  I wanted to test functionality not actual speed.
>>
>> All software was latest as of last night and I followed the quick how 
>> to here https://openvz.org/Parallels_Cloud_Storage
>>
>> Everything works great until I try to create an instance using the 
>> command "vzctl create 101 --layout ploop --ostemplate centos-6-x86_64 
>> --private /pcs/containers/101" from the docs.
>>
>> About one mb of data is written to disk and then it just hangs.  The 
>> following is output from dmesg
>>
>> [  360.414242] INFO: task vzctl:1646 blocked for more than 120 seconds.
>> [  360.414770] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" 
>> disables this message.
>> [  360.415406] vzctl         D ffff88007e444500     0 1646   1611    
>> 0 0x00000084
>> [  360.415418]  ffff88007ea59a68 0000000000000086 ffff8800ffffffff 
>> 000006b62934b8c0
>> [  360.415428]  0000000000000000 ffff88007e9f2ad0 0000000000005eaa 
>> ffffffffad17694d
>> [  360.415437]  000000000ad7ef74 ffffffff81a97b40 ffff88007e444ac8 
>> 000000000001eb80
>> [  360.415452] Call Trace:
>> [  360.415492]  [<ffffffff81517353>] io_schedule+0x73/0xc0
>> [  360.415516]  [<ffffffff811f39b3>] wait_on_sync_kiocb+0x53/0x80
>> [  360.415537]  [<ffffffffa04dbf47>] fuse_direct_IO+0x167/0x230 [fuse]
>> [  360.415558]  [<ffffffff8112e948>] mapping_direct_IO+0x48/0x70
>> [  360.415567]  [<ffffffff811301a6>] 
>> generic_file_direct_write_iter+0xf6/0x170
>> [  360.415576]  [<ffffffff81130c8e>] 
>> __generic_file_write_iter+0x32e/0x420
>> [  360.415585]  [<ffffffff81130e05>] __generic_file_aio_write+0x85/0xa0
>> [  360.415594]  [<ffffffff81130ea8>] generic_file_aio_write+0x88/0x100
>> [  360.415605]  [<ffffffffa04da085>] fuse_file_aio_write+0x185/0x430 
>> [fuse]
>> [  360.415623]  [<ffffffff811a530a>] do_sync_write+0xfa/0x140
>> [  360.415641]  [<ffffffff8109d930>] ? autoremove_wake_function+0x0/0x40
>> [  360.415655]  [<ffffffff812902da>] ? strncpy_from_user+0x4a/0x90
>> [  360.415664]  [<ffffffff811a55e8>] vfs_write+0xb8/0x1a0
>> [  360.415671]  [<ffffffff811a5ee1>] sys_write+0x51/0x90
>> [  360.415681]  [<ffffffff8100b102>] system_call_fastpath+0x16/0x1b
>>
>> Even just trying to create a 10k file with dd causes a task to hang.  
>> "dd if=/dev/zero of=/pcs/test.junk bs=1k count=10"
>>
>>
>> Any ideas? Anymore info you would like for debugging.
>> _______________________________________________
>> Users mailing list
>> Users at openvz.org <mailto:Users at openvz.org>
>> https://lists.openvz.org/mailman/listinfo/users
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openvz.org/pipermail/users/attachments/20140129/a501135b/attachment-0001.html>


More information about the Users mailing list