Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Greg Martyn via talk
I haven't used Gluster personally, but have you tried
turning performance.parallel-readdir on?
https://docs.gluster.org/en/latest/release-notes/3.10.0/#implemented-parallel-readdirp-with-distribute-xlator

It seems there's a reason why it's on by default (
https://www.spinics.net/lists/gluster-devel/msg25518.html) but maybe it'd
still be worth it for you?

On Mon, May 4, 2020 at 9:55 AM Alvin Starr via talk  wrote:

>
> I am hoping someone has seen this kind of problem before and knows of a
> solution.
> I have a client who has file systems filled with lots of small files on
> the orders of hundreds of millions of files.
> Running something like a find on filesystem takes the better part of a
> week so any kind of directory walking backup tool will take even longer
> to run.
> The actual data-size for 100M files is on the order of 15TB so there is
> a lot of data to backup but the data only increases on the order of tens
> to hundreds of MB a day.
>
>
> Even things like xfsdump take a long time.
> For example I tried xfsdump on a 50M file set and it took over 2 days to
> complete.
>
> The only thing that seems to be workable is Veeam.
> It will run an incremental volume snapshot in a few hours a night but I
> dislike adding proprietary kernel modules into the systems.
>
>
> --
> Alvin Starr   ||   land:  (647)478-6285
> Netvel Inc.   ||   Cell:  (416)806-0133
> al...@netvel.net  ||
>
> ---
> Post to this mailing list talk@gtalug.org
> Unsubscribe from this mailing list
> https://gtalug.org/mailman/listinfo/talk
>
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread John Sellens via talk
On Wed, 2020/05/06 10:38:29AM -0400, Howard Gibson via talk  
wrote:
| > ZFS is another option. And it handles delta-backups very easily.
| 
|How do you recover stuff from delta backups?  You have to figure which 
backup the file or directory is in, right?

Remember that snapshots, like RAID, are not actually backups,
unless they are on a different machine, in a different place.

ZFS makes it easy:

You can browse through snapshots for /mypool/myfs by looking
in /mypool/myfs/.zfs/snapshot and if your ZFS snapshots are
named using dates, easy peasy to choose when.  You can also
brute force and
  find /mypool/myfs/.zfs/snapshot -name 'myfile.tex' -ls
and find what's there.

You can use "zfs rollback" to revert to a snapshot.

You can use "zfs send ... | zfs recv ..." to copy a specific
snapshot (or group of snapshots) to another pool, system, etc.

And of course, when you create a snapshot, you could create your
own index listing of what's there for easy grepping.

ZFS is great.

You can still (and likely should) continue to backup to Blu-ray,
but ZFS will make sure your files don't rot in place unnoticed.

Hope that helps, cheers

John
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Alvin Starr via talk

On 5/6/20 10:18 AM, Lennart Sorensen via talk wrote:

On Wed, May 06, 2020 at 07:25:29AM -0400, David Mason wrote:

ZFS is another option. And it handles delta-backups very easily.

It is however not the XFS that glusterfs says to use.  If glusterfs is
involved, XFS really seems to be the only option.

There is Gluster hosted documentation on how to use ZFS with Gluster but 
it not the preferred option.

Putting Gluster on a thinlvm is the Gluster way for doing snapshots.

I have found a package called 
wyng-backup(https://github.com/tasket/wyng-backup).


This may solve my problem but it would require me to change all my 
volumes to thin provisioning.




--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Howard Gibson via talk
On Wed, 6 May 2020 07:25:29 -0400
David Mason via talk  wrote:

> ZFS is another option. And it handles delta-backups very easily.

David,

   How do you recover stuff from delta backups?  You have to figure which 
backup the file or directory is in, right?

   My backup recoveries, admittedly here at home, consist of me recovering 
files I stupidly modified incorrectly, and some accidentally deleted 
directories.  In the case of the directories, I noticed the problem several 
years after I did it.  Since I back up everything, all I had to do was pull out 
the backup Blu-ray from shortly after I worked on the directory. 

-- 
Howard Gibson 
hgib...@eol.ca
jhowardgib...@gmail.com
http://home.eol.ca/~hgibson
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Lennart Sorensen via talk
On Wed, May 06, 2020 at 07:25:29AM -0400, David Mason wrote:
> ZFS is another option. And it handles delta-backups very easily.

It is however not the XFS that glusterfs says to use.  If glusterfs is
involved, XFS really seems to be the only option.

-- 
Len Sorensen
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Alvin Starr via talk

On 5/6/20 7:25 AM, David Mason via talk wrote:

ZFS is another option. And it handles delta-backups very easily.

../Dave

I have been following ZFS on and off and it looks interesting but I am 
kind of stuck because it is not included in Centos/RH which is a 
requirement in this case.


It would be interesting to test it at scale just to see how well is does 
work.


--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread David Mason via talk
ZFS is another option. And it handles delta-backups very easily.

../Dave
On May 5, 2020, 11:27 PM -0400, Lennart Sorensen via talk , 
wrote:
> On Mon, May 04, 2020 at 10:42:25PM -0400, Alvin Starr via talk wrote:
> > The files are generally a few hundred KB each. They may run into a few MB
> > but that's about it.
> > I use to use ReiserFS back in the days of ext2/3 but it kind of fell out of
> > favor after the lead developer got sent away for murder.
> > Reiser was much faster and more reliable than ext at the time.
> > It would actually be interesting to see if running a reiserfs or btrfs
> > filesystem would actually make a significant difference but in the long run
> > I am kind of stuck with Centos/RH supported file systems and reiser and
> > btrfs are not part of that mix anymore.
>
> ReiserFS was not reliable. I certainly stopped using it long before
> the developer issues happened. The silent file content loss was just
> unacceptable. And it wasn't a rare occurance. I saw it on many systems
> many times. ext2 and 3 you could at least trust with your data even if
> they were quite a bit slower. Certainly these days RHEL supports ext2/3/4
> and XFS (their default and preferred). I use ext4 because it works well.
> GlusterFS defaults to XFS and while technically it can use other
> filesystems (and many people do run ext4 on it apparently) I don't
> believe they support that setup.
>
> > I am not sure how much I can get by tweaking the filesystem.
> > I would need to get a 50x -100x improvement to make backups complete in a
> > few hours.
> > Most stuff I have read comparing various filesystems and performance are
> > talking about percentage differences that is much less than 100%.
> >
> > I have a feeling that the only answer will be something like Veeam where
> > only changed blocks are backed up.
> > A directory tree walk just takes too long.
>
> Well, does the system have enough ram? That is something that often
> isn't hard to increase. XFS has certainly in the past been known to
> require a fair bit of ram to manage well.
>
> --
> Len Sorensen
> ---
> Post to this mailing list talk@gtalug.org
> Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Alvin Starr via talk

On 5/6/20 12:37 AM, Nicholas Krause wrote:
[snip]

Well, does the system have enough ram?  That is something that often

isn't hard to increase.  XFS has certainly in the past been known to
require a fair bit of ram to manage well.

I mentioned to check /proc/meminfo as well to look at cache in case you
missed that.

You did mention that and it is on my plate to research the amount of ram 
that is suggested for this kind of application.


--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-06 Thread Alvin Starr via talk

On 5/5/20 11:27 PM, Lennart Sorensen wrote:

On Mon, May 04, 2020 at 10:42:25PM -0400, Alvin Starr via talk wrote:

The files are generally a few hundred KB each. They may run into a few MB
but that's about it.
I use to use ReiserFS back in the days of ext2/3 but it kind of fell out of
favor after the lead developer got sent away for murder.
Reiser was much faster and more reliable than ext at the time.
It would actually be interesting to see if running a reiserfs or btrfs
filesystem would actually make a significant difference but in the long run
I am kind of stuck with Centos/RH supported file systems and reiser and
btrfs are not part of that mix anymore.

ReiserFS was not reliable.  I certainly stopped using it long before
the developer issues happened.  The silent file content loss was just
unacceptable.  And it wasn't a rare occurance.  I saw it on many systems
many times.  ext2 and 3 you could at least trust with your data even if
they were quite a bit slower.  Certainly these days RHEL supports ext2/3/4
and XFS (their default and preferred).  I use ext4 because it works well.
GlusterFS defaults to XFS and while technically it can use other
filesystems (and many people do run ext4 on it apparently) I don't
believe they support that setup.


This is a nice example of YMMV.
I moved my critical data from ext2/3 because they would trash my data 
and corrupt the whole filesystem under certain cases.
I had much better luck with Reiser even in the face of a couple of 
system crashes.

I am not sure how much I can get by tweaking the filesystem.
I would need to get a 50x -100x improvement to make backups complete in a
few hours.
Most stuff I have read comparing various filesystems and performance are
talking about percentage differences that is much less than 100%.

I have a feeling that the only answer will be something like Veeam where
only changed blocks are backed up.
A directory tree walk just takes too long.

Well, does the system have enough ram?  That is something that often
isn't hard to increase.  XFS has certainly in the past been known to
require a fair bit of ram to manage well.


The system has 32G and runs nothing but gluster.
Not sure if that is enough ram and will require a bit of researching.
It looks like the system has about 15G of cached data.

--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-05 Thread Nicholas Krause via talk



On 5/5/20 11:27 PM, Lennart Sorensen wrote:

On Mon, May 04, 2020 at 10:42:25PM -0400, Alvin Starr via talk wrote:

The files are generally a few hundred KB each. They may run into a few MB
but that's about it.
I use to use ReiserFS back in the days of ext2/3 but it kind of fell out of
favor after the lead developer got sent away for murder.
Reiser was much faster and more reliable than ext at the time.
It would actually be interesting to see if running a reiserfs or btrfs
filesystem would actually make a significant difference but in the long run
I am kind of stuck with Centos/RH supported file systems and reiser and
btrfs are not part of that mix anymore.

ReiserFS was not reliable.  I certainly stopped using it long before
the developer issues happened.  The silent file content loss was just
unacceptable.  And it wasn't a rare occurance.  I saw it on many systems
many times.  ext2 and 3 you could at least trust with your data even if
they were quite a bit slower.  Certainly these days RHEL supports ext2/3/4
and XFS (their default and preferred).  I use ext4 because it works well.
GlusterFS defaults to XFS and while technically it can use other
filesystems (and many people do run ext4 on it apparently) I don't
believe they support that setup.


I am not sure how much I can get by tweaking the filesystem.
I would need to get a 50x -100x improvement to make backups complete in a
few hours.
Most stuff I have read comparing various filesystems and performance are
talking about percentage differences that is much less than 100%.

I have a feeling that the only answer will be something like Veeam where
only changed blocks are backed up.
A directory tree walk just takes too long.

Well, does the system have enough ram?  That is something that often
isn't hard to increase.  XFS has certainly in the past been known to
require a fair bit of ram to manage well.

I mentioned to check /proc/meminfo as well to look at cache in case you
missed that.

Nick




--
Fundamentally an organism has conscious mental states if and only if there is 
something that it is like to be that organism--something it is like for the 
organism. - Thomas Nagel

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-05 Thread Lennart Sorensen via talk
On Mon, May 04, 2020 at 10:42:25PM -0400, Alvin Starr via talk wrote:
> The files are generally a few hundred KB each. They may run into a few MB
> but that's about it.
> I use to use ReiserFS back in the days of ext2/3 but it kind of fell out of
> favor after the lead developer got sent away for murder.
> Reiser was much faster and more reliable than ext at the time.
> It would actually be interesting to see if running a reiserfs or btrfs
> filesystem would actually make a significant difference but in the long run
> I am kind of stuck with Centos/RH supported file systems and reiser and
> btrfs are not part of that mix anymore.

ReiserFS was not reliable.  I certainly stopped using it long before
the developer issues happened.  The silent file content loss was just
unacceptable.  And it wasn't a rare occurance.  I saw it on many systems
many times.  ext2 and 3 you could at least trust with your data even if
they were quite a bit slower.  Certainly these days RHEL supports ext2/3/4
and XFS (their default and preferred).  I use ext4 because it works well.
GlusterFS defaults to XFS and while technically it can use other
filesystems (and many people do run ext4 on it apparently) I don't
believe they support that setup.

> I am not sure how much I can get by tweaking the filesystem.
> I would need to get a 50x -100x improvement to make backups complete in a
> few hours.
> Most stuff I have read comparing various filesystems and performance are
> talking about percentage differences that is much less than 100%.
> 
> I have a feeling that the only answer will be something like Veeam where
> only changed blocks are backed up.
> A directory tree walk just takes too long.

Well, does the system have enough ram?  That is something that often
isn't hard to increase.  XFS has certainly in the past been known to
require a fair bit of ram to manage well.

-- 
Len Sorensen
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Nicholas Krause via talk



On 5/4/20 10:42 PM, Alvin Starr wrote:

On 5/4/20 7:28 PM, nick wrote:

On 2020-05-04 2:12 p.m., Alvin Starr via talk wrote:

On 5/4/20 1:26 PM, Lennart Sorensen via talk wrote:

On Mon, May 04, 2020 at 04:38:28PM +0200, ac via talk wrote:

Hi Alvin,

On a 2TB dataset, with +-600k files, I have piped tree to less with
limited joy, it took a few hours and at least I could search for
what I was looking for... - 15TB and 100M is another animal though
and as disk i/o will be your bottleneck, anything will take long, no?

now, for my own info/interest, can you tell me which fs is used 
for this

ext3?

Hmm, sounds awful slow.

Just for fun I ran find on one of my drives:

# time find /data | wc -l
1825463
real    3m57s.208s

That is with 5.3T used out of 6.0TB.

Running it a second time when it is cached takes 7.7s.  Tree takes 
14.7s.


Another volume:
# time find /mythdata | wc -l
54972

real    0m1.924s

That is with 15 TB out of 15 TB in use (yes that one always fills up
for some reason).

Both of those are lvm volumes with ext4 on top of software raid6 using
5400rpm WD red drives.

Seems either XFS is unbelievable bad, or there isn't enough ram to 
cache

the filesystem metadata if you are having a problem with 100M files.
I only have a measly 32GB in my home machine.

I believe the directory hierarchy has a lot to do with the performance.
It seems that the listing time is non-linear although I do not 
believe its  an N^^2 kind of problem.
I would have said the same as you before I started having to deal 
with 10's of millions of files.




The first question I would have is how big are the actual files 
versus space used. Most file systems
try to merge files that don't allocate in a single block to save 
blocks and get better space usage.
Assuming a) the size of the files being rather small and b) The 
amount, I would be curious if ReiserFS
or brtfs helps either as most have much better tail merging into 
metadata blocks to my knowledge. Better
packing the small files can help as disk seeks are a problem here it 
seems. My disks will solve to like

10 MB/S on ext4 with lots of small files like this due to seeks.
The files are generally a few hundred KB each. They may run into a few 
MB but that's about it.
I use to use ReiserFS back in the days of ext2/3 but it kind of fell 
out of favor after the lead developer got sent away for murder.

Reiser was much faster and more reliable than ext at the time.
It would actually be interesting to see if running a reiserfs or btrfs 
filesystem would actually make a significant difference but in the 
long run I am kind of stuck with Centos/RH supported file systems and 
reiser and btrfs are not part of that mix anymore.
Interesting, I recall CentOS 7 and RHEL 7 supporting it. You must be on 
a older system unless my memory is wrong :). The problem really is
seek times. I've not sure if there is a way to tune the kernel from 
memory for this but that's your problem unfortunately.


I am not sure how much I can get by tweaking the filesystem.
I would need to get a 50x -100x improvement to make backups complete 
in a few hours.
Most stuff I have read comparing various filesystems and performance 
are talking about percentage differences that is much less than 100%.


I have a feeling that the only answer will be something like Veeam 
where only changed blocks are backed up.

A directory tree walk just takes too long.
See my above comments but the other two ideas would be to use some short 
of filesystem or LVM snapshots
or  try dd as a benchmark to see if a raw copy tool does better. 
Snapshots are probably the best way through
or only changed blocked based tools as you logically came to the 
conclusion of.


Good luck through as lots of small files are always a pain in 
transferring and especially on rotating disks,

Nick




The other question as pointed out here is how much memory the page 
cache and other kernel caches are
using. I would check /proc/meminfo to start as this may be another 
logical solution already pointed

out.

Maybe that helps,

Nick




--
Fundamentally an organism has conscious mental states if and only if there is 
something that it is like to be that organism--something it is like for the 
organism. - Thomas Nagel

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Alvin Starr via talk

On 5/4/20 7:28 PM, nick wrote:

On 2020-05-04 2:12 p.m., Alvin Starr via talk wrote:

On 5/4/20 1:26 PM, Lennart Sorensen via talk wrote:

On Mon, May 04, 2020 at 04:38:28PM +0200, ac via talk wrote:

Hi Alvin,

On a 2TB dataset, with +-600k files, I have piped tree to less with
limited joy, it took a few hours and at least I could search for
what I was looking for... - 15TB and 100M is another animal though
and as disk i/o will be your bottleneck, anything will take long, no?

now, for my own info/interest, can you tell me which fs is used for this
ext3?

Hmm, sounds awful slow.

Just for fun I ran find on one of my drives:

# time find /data | wc -l
1825463
real    3m57s.208s

That is with 5.3T used out of 6.0TB.

Running it a second time when it is cached takes 7.7s.  Tree takes 14.7s.

Another volume:
# time find /mythdata | wc -l
54972

real    0m1.924s

That is with 15 TB out of 15 TB in use (yes that one always fills up
for some reason).

Both of those are lvm volumes with ext4 on top of software raid6 using
5400rpm WD red drives.

Seems either XFS is unbelievable bad, or there isn't enough ram to cache
the filesystem metadata if you are having a problem with 100M files.
I only have a measly 32GB in my home machine.

I believe the directory hierarchy has a lot to do with the performance.
It seems that the listing time is non-linear although I do not believe its  an 
N^^2 kind of problem.
I would have said the same as you before I started having to deal with 10's of 
millions of files.




The first question I would have is how big are the actual files versus space 
used. Most file systems
try to merge files that don't allocate in a single block to save blocks and get 
better space usage.
Assuming a) the size of the files being rather small and b) The amount, I would 
be curious if ReiserFS
or brtfs helps either as most have much better tail merging into metadata 
blocks to my knowledge. Better
packing the small files can help as disk seeks are a problem here it seems. My 
disks will solve to like
10 MB/S on ext4 with lots of small files like this due to seeks.
The files are generally a few hundred KB each. They may run into a few 
MB but that's about it.
I use to use ReiserFS back in the days of ext2/3 but it kind of fell out 
of favor after the lead developer got sent away for murder.

Reiser was much faster and more reliable than ext at the time.
It would actually be interesting to see if running a reiserfs or btrfs 
filesystem would actually make a significant difference but in the long 
run I am kind of stuck with Centos/RH supported file systems and reiser 
and btrfs are not part of that mix anymore.


I am not sure how much I can get by tweaking the filesystem.
I would need to get a 50x -100x improvement to make backups complete in 
a few hours.
Most stuff I have read comparing various filesystems and performance are 
talking about percentage differences that is much less than 100%.


I have a feeling that the only answer will be something like Veeam where 
only changed blocks are backed up.

A directory tree walk just takes too long.



The other question as pointed out here is how much memory the page cache and 
other kernel caches are
using. I would check /proc/meminfo to start as this may be another logical 
solution already pointed
out.

Maybe that helps,

Nick


--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread nick via talk


On 2020-05-04 2:12 p.m., Alvin Starr via talk wrote:
> On 5/4/20 1:26 PM, Lennart Sorensen via talk wrote:
>> On Mon, May 04, 2020 at 04:38:28PM +0200, ac via talk wrote:
>>> Hi Alvin,
>>>
>>> On a 2TB dataset, with +-600k files, I have piped tree to less with
>>> limited joy, it took a few hours and at least I could search for
>>> what I was looking for... - 15TB and 100M is another animal though
>>> and as disk i/o will be your bottleneck, anything will take long, no?
>>>
>>> now, for my own info/interest, can you tell me which fs is used for this
>>> ext3?
>> Hmm, sounds awful slow.
>>
>> Just for fun I ran find on one of my drives:
>>
>> # time find /data | wc -l
>> 1825463
>> real    3m57s.208s
>>
>> That is with 5.3T used out of 6.0TB.
>>
>> Running it a second time when it is cached takes 7.7s.  Tree takes 14.7s.
>>
>> Another volume:
>> # time find /mythdata | wc -l
>> 54972
>>
>> real    0m1.924s
>>
>> That is with 15 TB out of 15 TB in use (yes that one always fills up
>> for some reason).
>>
>> Both of those are lvm volumes with ext4 on top of software raid6 using
>> 5400rpm WD red drives.
>>
>> Seems either XFS is unbelievable bad, or there isn't enough ram to cache
>> the filesystem metadata if you are having a problem with 100M files.
>> I only have a measly 32GB in my home machine.
> 
> I believe the directory hierarchy has a lot to do with the performance.
> It seems that the listing time is non-linear although I do not believe its  
> an N^^2 kind of problem.
> I would have said the same as you before I started having to deal with 10's 
> of millions of files.
> 
> 
> 

The first question I would have is how big are the actual files versus space 
used. Most file systems
try to merge files that don't allocate in a single block to save blocks and get 
better space usage.
Assuming a) the size of the files being rather small and b) The amount, I would 
be curious if ReiserFS 
or brtfs helps either as most have much better tail merging into metadata 
blocks to my knowledge. Better
packing the small files can help as disk seeks are a problem here it seems. My 
disks will solve to like
10 MB/S on ext4 with lots of small files like this due to seeks.

The other question as pointed out here is how much memory the page cache and 
other kernel caches are
using. I would check /proc/meminfo to start as this may be another logical 
solution already pointed
out.

Maybe that helps,

Nick
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Alvin Starr via talk

On 5/4/20 1:26 PM, Lennart Sorensen via talk wrote:

On Mon, May 04, 2020 at 04:38:28PM +0200, ac via talk wrote:

Hi Alvin,

On a 2TB dataset, with +-600k files, I have piped tree to less with
limited joy, it took a few hours and at least I could search for
what I was looking for... - 15TB and 100M is another animal though
and as disk i/o will be your bottleneck, anything will take long, no?

now, for my own info/interest, can you tell me which fs is used for this
ext3?

Hmm, sounds awful slow.

Just for fun I ran find on one of my drives:

# time find /data | wc -l
1825463
real3m57s.208s

That is with 5.3T used out of 6.0TB.

Running it a second time when it is cached takes 7.7s.  Tree takes 14.7s.

Another volume:
# time find /mythdata | wc -l
54972

real0m1.924s

That is with 15 TB out of 15 TB in use (yes that one always fills up
for some reason).

Both of those are lvm volumes with ext4 on top of software raid6 using
5400rpm WD red drives.

Seems either XFS is unbelievable bad, or there isn't enough ram to cache
the filesystem metadata if you are having a problem with 100M files.
I only have a measly 32GB in my home machine.


I believe the directory hierarchy has a lot to do with the performance.
It seems that the listing time is non-linear although I do not believe 
its  an N^^2 kind of problem.
I would have said the same as you before I started having to deal with 
10's of millions of files.




--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Alvin Starr via talk

On 5/4/20 12:52 PM, John Sellens wrote:

On Mon, 2020/05/04 12:03:19PM -0400, Alvin Starr  wrote:
| The client really only wants to use Centos/RHEL and ZFS is not part of that
| mix at the moment.

Well, one could argue that zfs on centos is fairly well supported ...
If it were purely my choice I would agree but the general rule is 
software that can be gotten from Centos/RH and EPEL where necessary.




| The data is actually sitting on a replicated Gluster cluster so trying to
| replace that with an HA NAS would start to get expensive if it were a
| commercial product.

Of course "expensive" depends on the client.  An HA truenas that size,
all flash is (I believe likely well) under $15K USD.
If  things moved to a commercial NAS it would likely be something like 
Equalogic or Netapp and in that world if you have to ask the price you 
cannot afford it.



Ah - you didn't mention Gluster.

In theory, Gluster has geographic replication.

It does have replication but replication is not backup.
a little oops like "rm -rf * somefile"  will make for a very bad day if 
you don't catch it before it gets replicated.




And if your bricks are on LVM storage, you can use gluster snapshots as well:
 
https://docs.gluster.org/en/latest/Administrator%20Guide/Managing%20Snapshots/
to guard against accidental removals, etc.

Gluster snapshots require thinly provisioned LVM.
Which is doable  but will require rebuilding the systems with a 
different LVM config.


(I've not used either, and my glusters are quite old versions at the present 
time.)

Depending on how it's all configured, you may get better performance
backing up the bricks, rather than backing up gluster itself.  I have
a two-node gluster, mirrored, so I can backup the bricks on one of the
servers and get everything.  Obviously that's a very simple "cluster".

Traditionally, gluster filesystem performance with large numbers of
small files in a directory is horrible/pathetic.  If you're backing up
the gluster filesystem, you would almost certainly get better performance
if your file structure is deeper and narrower, if that's possible.

We are actually backing up the bricks but the they are BIG bricks.
Yes gluster has a high per file open cost but in this application that 
is not an issue during operation.
A backup using a gluster mount would move from the world of days to 
weeks because of the synchronization overhead.


Even with snapshots the length of time for a backup can be outrageously 
long.


Cheers

John


--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Lennart Sorensen via talk
On Mon, May 04, 2020 at 04:38:28PM +0200, ac via talk wrote:
> Hi Alvin,
> 
> On a 2TB dataset, with +-600k files, I have piped tree to less with 
> limited joy, it took a few hours and at least I could search for
> what I was looking for... - 15TB and 100M is another animal though
> and as disk i/o will be your bottleneck, anything will take long, no?
> 
> now, for my own info/interest, can you tell me which fs is used for this
> ext3? 

Hmm, sounds awful slow.

Just for fun I ran find on one of my drives:

# time find /data | wc -l
1825463
real3m57s.208s

That is with 5.3T used out of 6.0TB.

Running it a second time when it is cached takes 7.7s.  Tree takes 14.7s.

Another volume:
# time find /mythdata | wc -l
54972

real0m1.924s

That is with 15 TB out of 15 TB in use (yes that one always fills up
for some reason).

Both of those are lvm volumes with ext4 on top of software raid6 using
5400rpm WD red drives.

Seems either XFS is unbelievable bad, or there isn't enough ram to cache
the filesystem metadata if you are having a problem with 100M files.
I only have a measly 32GB in my home machine.

-- 
Len Sorensen
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread John Sellens via talk
On Mon, 2020/05/04 12:03:19PM -0400, Alvin Starr  wrote:
| The client really only wants to use Centos/RHEL and ZFS is not part of that
| mix at the moment.

Well, one could argue that zfs on centos is fairly well supported ...


| The data is actually sitting on a replicated Gluster cluster so trying to
| replace that with an HA NAS would start to get expensive if it were a
| commercial product.

Of course "expensive" depends on the client.  An HA truenas that size,
all flash is (I believe likely well) under $15K USD.


Ah - you didn't mention Gluster.

In theory, Gluster has geographic replication.

And if your bricks are on LVM storage, you can use gluster snapshots as well:

https://docs.gluster.org/en/latest/Administrator%20Guide/Managing%20Snapshots/
to guard against accidental removals, etc.

(I've not used either, and my glusters are quite old versions at the present 
time.)

Depending on how it's all configured, you may get better performance
backing up the bricks, rather than backing up gluster itself.  I have
a two-node gluster, mirrored, so I can backup the bricks on one of the
servers and get everything.  Obviously that's a very simple "cluster".

Traditionally, gluster filesystem performance with large numbers of
small files in a directory is horrible/pathetic.  If you're backing up
the gluster filesystem, you would almost certainly get better performance
if your file structure is deeper and narrower, if that's possible.

Cheers

John
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Jamon Camisso via talk
On 2020-05-04 09:55, Alvin Starr via talk wrote:
> 
> I am hoping someone has seen this kind of problem before and knows of a
> solution.
> I have a client who has file systems filled with lots of small files on
> the orders of hundreds of millions of files.
> Running something like a find on filesystem takes the better part of a
> week so any kind of directory walking backup tool will take even longer
> to run.
> The actual data-size for 100M files is on the order of 15TB so there is
> a lot of data to backup but the data only increases on the order of tens
> to hundreds of MB a day.
> 
> 
> Even things like xfsdump take a long time.
> For example I tried xfsdump on a 50M file set and it took over 2 days to
> complete.
> 
> The only thing that seems to be workable is Veeam.
> It will run an incremental volume snapshot in a few hours a night but I
> dislike adding proprietary kernel modules into the systems.

If you have a list of inodes on the filesystem you can use xfs_db directly:

xfs_db> inode 128
xfs_db> blockget -n
xfs_db> ncheck
131 dir/.
132 dir/test2/foo/.
133 dir/test2/foo/bar
  65664 dir/test1/.
  65665 dir/test1/foo
  65666 dir/test3/foo/.
 142144 dir/test2/.
 142145 dir/test3/foo/bar/.
 142146 dir/test3/foo/bar/baz
 196736 dir/test3/.

I don't know how that will perform relative to something like find though.

Cheers, Jamon
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Alvin Starr via talk

Sadly this one is a bit of a non-starter.

The client really only wants to use Centos/RHEL and ZFS is not part of 
that mix at the moment.


The data is actually sitting on a replicated Gluster cluster so trying 
to replace that with an HA NAS would start to get expensive if it were a 
commercial product.


On 5/4/20 11:27 AM, John Sellens wrote:

I bet no one would want this advice, but it seems to me that the
implementation needs to change i.e. that one big (possibly shallow)
filesystem on xfs is unworkable.

The best answer of course depends on the value of the data.

One obvious approach is to use a filesystem/NAS with off-site replication.
Typically a commerical product.

At relatively modest cost, I like the truenas systems from ixsystems.com.
ZFS based, HA versions available, replication can be done.
The HA versions are two servers in one chassis, with dual-ported SAS disks.

For do-it-yourselfers, I like using ZFS and ZFS replication of snapshots.
For example, I do much (much) smaller offsites from my home to work
using ZFS and zfs-replicate.

You can also do freenas (non-commercial truenas) but without the HA
hardware and code.

Hope that helps - cheers

John


On Mon, 2020/05/04 09:55:51AM -0400, Alvin Starr via talk  
wrote:
| The actual data-size for 100M files is on the order of 15TB so there is a
| lot of data to backup but the data only increases on the order of tens to
| hundreds of MB a day.


--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread John Sellens via talk
I bet no one would want this advice, but it seems to me that the
implementation needs to change i.e. that one big (possibly shallow)
filesystem on xfs is unworkable.

The best answer of course depends on the value of the data.

One obvious approach is to use a filesystem/NAS with off-site replication.
Typically a commerical product.

At relatively modest cost, I like the truenas systems from ixsystems.com.
ZFS based, HA versions available, replication can be done.
The HA versions are two servers in one chassis, with dual-ported SAS disks.

For do-it-yourselfers, I like using ZFS and ZFS replication of snapshots.
For example, I do much (much) smaller offsites from my home to work
using ZFS and zfs-replicate.

You can also do freenas (non-commercial truenas) but without the HA
hardware and code.

Hope that helps - cheers

John


On Mon, 2020/05/04 09:55:51AM -0400, Alvin Starr via talk  
wrote:
| The actual data-size for 100M files is on the order of 15TB so there is a
| lot of data to backup but the data only increases on the order of tens to
| hundreds of MB a day.
---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread Alvin Starr via talk
I am not quite sure where the breaking point is but I think part of the 
problem is that the directories start to get big.

The directory hierarchy is only 5 to 10 nodes deep.

Its running on xfs.


On 5/4/20 10:38 AM, ac wrote:

Hi Alvin,

On a 2TB dataset, with +-600k files, I have piped tree to less with
limited joy, it took a few hours and at least I could search for
what I was looking for... - 15TB and 100M is another animal though
and as disk i/o will be your bottleneck, anything will take long, no?

now, for my own info/interest, can you tell me which fs is used for this
ext3?

On Mon, 4 May 2020 09:55:51 -0400
Alvin Starr via talk  wrote:

I am hoping someone has seen this kind of problem before and knows of
a solution.
I have a client who has file systems filled with lots of small files
on the orders of hundreds of millions of files.
Running something like a find on filesystem takes the better part of
a week so any kind of directory walking backup tool will take even
longer to run.
The actual data-size for 100M files is on the order of 15TB so there
is a lot of data to backup but the data only increases on the order
of tens to hundreds of MB a day.


Even things like xfsdump take a long time.
For example I tried xfsdump on a 50M file set and it took over 2 days
to complete.

The only thing that seems to be workable is Veeam.
It will run an incremental volume snapshot in a few hours a night but
I dislike adding proprietary kernel modules into the systems.




--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


Re: [GTALUG] On the subject of backups.

2020-05-04 Thread ac via talk
Hi Alvin,

On a 2TB dataset, with +-600k files, I have piped tree to less with 
limited joy, it took a few hours and at least I could search for
what I was looking for... - 15TB and 100M is another animal though
and as disk i/o will be your bottleneck, anything will take long, no?

now, for my own info/interest, can you tell me which fs is used for this
ext3? 

On Mon, 4 May 2020 09:55:51 -0400
Alvin Starr via talk  wrote:
> I am hoping someone has seen this kind of problem before and knows of
> a solution.
> I have a client who has file systems filled with lots of small files
> on the orders of hundreds of millions of files.
> Running something like a find on filesystem takes the better part of
> a week so any kind of directory walking backup tool will take even
> longer to run.
> The actual data-size for 100M files is on the order of 15TB so there
> is a lot of data to backup but the data only increases on the order
> of tens to hundreds of MB a day.
> 
> 
> Even things like xfsdump take a long time.
> For example I tried xfsdump on a 50M file set and it took over 2 days
> to complete.
> 
> The only thing that seems to be workable is Veeam.
> It will run an incremental volume snapshot in a few hours a night but
> I dislike adding proprietary kernel modules into the systems.
> 
> 

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk


[GTALUG] On the subject of backups.

2020-05-04 Thread Alvin Starr via talk


I am hoping someone has seen this kind of problem before and knows of a 
solution.
I have a client who has file systems filled with lots of small files on 
the orders of hundreds of millions of files.
Running something like a find on filesystem takes the better part of a 
week so any kind of directory walking backup tool will take even longer 
to run.
The actual data-size for 100M files is on the order of 15TB so there is 
a lot of data to backup but the data only increases on the order of tens 
to hundreds of MB a day.



Even things like xfsdump take a long time.
For example I tried xfsdump on a 50M file set and it took over 2 days to 
complete.


The only thing that seems to be workable is Veeam.
It will run an incremental volume snapshot in a few hours a night but I 
dislike adding proprietary kernel modules into the systems.



--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net  ||

---
Post to this mailing list talk@gtalug.org
Unsubscribe from this mailing list https://gtalug.org/mailman/listinfo/talk