Hi John,
That's great, thank you so much for the advice.
Some of our users have massive files so this would have been a big block.
Is there any particular reason for having a file size limit? Would setting
max_file_size to 0 remove all limits?
Thanks again,
Jake
On 24 May 2017 19:45:52 BST, John Spray <[email protected]> wrote:
>On Wed, May 24, 2017 at 7:41 PM, Brady Deetz <[email protected]> wrote:
>> Are there any repercussions to configuring this on an existing large
>fs?
>
>No. It's just a limit that's enforced at the point of appending to
>files or setting their size, it doesn't affect how anything is stored.
>
>John
>
>> On Wed, May 24, 2017 at 1:36 PM, John Spray <[email protected]>
>wrote:
>>>
>>> On Wed, May 24, 2017 at 7:19 PM, Jake Grimmett
><[email protected]>
>>> wrote:
>>> > Dear All,
>>> >
>>> > I've been testing out cephfs, and bumped into what appears to be
>an
>>> > upper
>>> > file size limit of ~1.1TB
>>> >
>>> > e.g:
>>> >
>>> > [root@cephfs1 ~]# time rsync --progress -av /ssd/isilon_melis.tar
>>> > /ceph/isilon_melis.tar
>>> > sending incremental file list
>>> > isilon_melis.tar
>>> > 1099341824000 54% 237.51MB/s 1:02:05
>>> > rsync: writefd_unbuffered failed to write 4 bytes to socket
>[sender]:
>>> > Broken pipe (32)
>>> > rsync: write failed on "/ceph/isilon_melis.tar": File too large
>(27)
>>> > rsync error: error in file IO (code 11) at receiver.c(322)
>>> > [receiver=3.0.9]
>>> > rsync: connection unexpectedly closed (28 bytes received so far)
>>> > [sender]
>>> > rsync error: error in rsync protocol data stream (code 12) at
>io.c(605)
>>> > [sender=3.0.9]
>>> >
>>> > Firstly, is this expected?
>>>
>>> CephFS has a configurable maximum file size, it's 1TB by default.
>>>
>>> Change it with:
>>> ceph fs set <fs name> max_file_size <size in bytes>
>>>
>>> John
>>>
>>>
>>>
>>>
>>>
>>> >
>>> > If not, then does anyone have any suggestions on where to start
>digging?
>>> >
>>> > I'm using erasure encoding (4+1, 50 x 8TB drives over 5 servers),
>with
>>> > an
>>> > nvme hot pool of 4 drives (2 x replication).
>>> >
>>> > I've tried both Kraken (release), and the latest Luminous Dev.
>>> >
>>> > many thanks,
>>> >
>>> > Jake
>>> > --
>>> >
>>> > _______________________________________________
>>> > ceph-users mailing list
>>> > [email protected]
>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
--
Sent from my Android device with K-9 Mail. Please excuse my brevity.
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com