Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-14 Thread Dave Chinner
that shows in the runtime which also drops from 3m57s to 3m22s. So regardless of what aim7 results we get from these changes, I'll be merging them pending review and further testing... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-14 Thread Dave Chinner
that shows in the runtime which also drops from 3m57s to 3m22s. So regardless of what aim7 results we get from these changes, I'll be merging them pending review and further testing... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 3.16 102/305] xfs: xfs_iflush_cluster fails to abort on error

2016-08-14 Thread Dave Chinner
On Sat, Aug 13, 2016 at 06:42:51PM +0100, Ben Hutchings wrote: > 3.16.37-rc1 review patch. If anyone has any objections, please let me know. > > -- > > From: Dave Chinner <dchin...@redhat.com> > > commit b1438f477934f5a4d5a44df26f3079a7575d5946 upstre

Re: [PATCH 3.16 102/305] xfs: xfs_iflush_cluster fails to abort on error

2016-08-14 Thread Dave Chinner
On Sat, Aug 13, 2016 at 06:42:51PM +0100, Ben Hutchings wrote: > 3.16.37-rc1 review patch. If anyone has any objections, please let me know. > > -- > > From: Dave Chinner > > commit b1438f477934f5a4d5a44df26f3079a7575d5946 upstream. > > When a fai

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-14 Thread Dave Chinner
On Sat, Aug 13, 2016 at 02:30:54AM +0200, Christoph Hellwig wrote: > On Fri, Aug 12, 2016 at 08:02:08PM +1000, Dave Chinner wrote: > > Which says "no change". Oh well, back to the drawing board... > > I don't see how it would change thing much - for all relevant calculati

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-14 Thread Dave Chinner
On Sat, Aug 13, 2016 at 02:30:54AM +0200, Christoph Hellwig wrote: > On Fri, Aug 12, 2016 at 08:02:08PM +1000, Dave Chinner wrote: > > Which says "no change". Oh well, back to the drawing board... > > I don't see how it would change thing much - for all relevant calculati

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-12 Thread Dave Chinner
On Fri, Aug 12, 2016 at 04:51:24PM +0800, Ye Xiaolong wrote: > On 08/12, Ye Xiaolong wrote: > >On 08/12, Dave Chinner wrote: > > [snip] > > >>lkp-folk: the patch I've just tested it attached below - can you > >>feed that through your test and see if it fixes

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-12 Thread Dave Chinner
On Fri, Aug 12, 2016 at 04:51:24PM +0800, Ye Xiaolong wrote: > On 08/12, Ye Xiaolong wrote: > >On 08/12, Dave Chinner wrote: > > [snip] > > >>lkp-folk: the patch I've just tested it attached below - can you > >>feed that through your test and see if it fixes

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-12 Thread Dave Chinner
On Thu, Aug 11, 2016 at 10:02:39PM -0700, Linus Torvalds wrote: > On Thu, Aug 11, 2016 at 9:16 PM, Dave Chinner <da...@fromorbit.com> wrote: > > > > That's why running aim7 as your "does the filesystem scale" > > benchmark is somewhat irrelevant to scaling

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-12 Thread Dave Chinner
On Thu, Aug 11, 2016 at 10:02:39PM -0700, Linus Torvalds wrote: > On Thu, Aug 11, 2016 at 9:16 PM, Dave Chinner wrote: > > > > That's why running aim7 as your "does the filesystem scale" > > benchmark is somewhat irrelevant to scaling applications on high >

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
's why running aim7 as your "does the filesystem scale" benchmark is somewhat irrelevant to scaling applications on high performance systems these days - users with fast storage will be expecting to see that 1.9GB/s throughput from their app, not 600MB/s Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
aim7 as your "does the filesystem scale" benchmark is somewhat irrelevant to scaling applications on high performance systems these days - users with fast storage will be expecting to see that 1.9GB/s throughput from their app, not 600MB/s Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
On Thu, Aug 11, 2016 at 07:27:52PM -0700, Linus Torvalds wrote: > On Thu, Aug 11, 2016 at 5:54 PM, Dave Chinner <da...@fromorbit.com> wrote: > > > > So, removing mark_page_accessed() made the spinlock contention > > *worse*. > > > > 36.51% [kernel] [k]

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
On Thu, Aug 11, 2016 at 07:27:52PM -0700, Linus Torvalds wrote: > On Thu, Aug 11, 2016 at 5:54 PM, Dave Chinner wrote: > > > > So, removing mark_page_accessed() made the spinlock contention > > *worse*. > > > > 36.51% [kernel] [k] _raw_spin_unlock_irqr

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
On Fri, Aug 12, 2016 at 10:54:42AM +1000, Dave Chinner wrote: > I'm now going to test Christoph's theory that this is an "overwrite > doing lots of block mapping" issue. More on that to follow. Ok, so going back to the profiles, I can say it's not an overwrite issue, because

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
On Fri, Aug 12, 2016 at 10:54:42AM +1000, Dave Chinner wrote: > I'm now going to test Christoph's theory that this is an "overwrite > doing lots of block mapping" issue. More on that to follow. Ok, so going back to the profiles, I can say it's not an overwrite issue, because

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
On Thu, Aug 11, 2016 at 11:16:12AM +1000, Dave Chinner wrote: > On Wed, Aug 10, 2016 at 05:33:20PM -0700, Huang, Ying wrote: > We need to know what is happening that is different - there's a good > chance the mapping trace events will tell us. Huang, can you get > a raw event trace f

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
On Thu, Aug 11, 2016 at 11:16:12AM +1000, Dave Chinner wrote: > On Wed, Aug 10, 2016 at 05:33:20PM -0700, Huang, Ying wrote: > We need to know what is happening that is different - there's a good > chance the mapping trace events will tell us. Huang, can you get > a raw event trace f

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
level - the mapping->tree_lock is a global serialisation point I'm now going to test Christoph's theory that this is an "overwrite doing lots of block mapping" issue. More on that to follow. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-11 Thread Dave Chinner
ping->tree_lock is a global serialisation point I'm now going to test Christoph's theory that this is an "overwrite doing lots of block mapping" issue. More on that to follow. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
nlock_irqrestore I don't think that this is the same as what aim7 is triggering as there's no XFS write() path allocation functions near the top of the profile to speak of. Still, I don't recall seeing this before... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
nk that this is the same as what aim7 is triggering as there's no XFS write() path allocation functions near the top of the profile to speak of. Still, I don't recall seeing this before... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
On Thu, Aug 11, 2016 at 10:36:59AM +0800, Ye Xiaolong wrote: > On 08/11, Dave Chinner wrote: > >On Thu, Aug 11, 2016 at 11:16:12AM +1000, Dave Chinner wrote: > >> I need to see these events: > >> > >>xfs_file* > >>xfs_iomap* > >>

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
On Thu, Aug 11, 2016 at 10:36:59AM +0800, Ye Xiaolong wrote: > On 08/11, Dave Chinner wrote: > >On Thu, Aug 11, 2016 at 11:16:12AM +1000, Dave Chinner wrote: > >> I need to see these events: > >> > >>xfs_file* > >>xfs_iomap* > >>

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
On Thu, Aug 11, 2016 at 11:16:12AM +1000, Dave Chinner wrote: > I need to see these events: > > xfs_file* > xfs_iomap* > xfs_get_block* > > For both kernels. An example trace from 4.8-rc1 running the command > `xfs_io -f -c 'pwrite 0 512k -b 128k'

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
On Thu, Aug 11, 2016 at 11:16:12AM +1000, Dave Chinner wrote: > I need to see these events: > > xfs_file* > xfs_iomap* > xfs_get_block* > > For both kernels. An example trace from 4.8-rc1 running the command > `xfs_io -f -c 'pwrite 0 512k -b 128k'

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
. 253971.751234: xfs_file_buffered_write: dev 253:32 ino 0x84 size 0x4 offset 0x4 count 0x2 xfs_io-2946 [001] 253971.751236: xfs_iomap_found: dev 253:32 ino 0x84 size 0x40000 offset 0x4 count 131072 type invalid startoff 0x0 startblock 24 blockcount 0x60 xfs_io-2946 [001] 253971.751381: xfs_file_buffered_write: dev 253:32 ino 0x84 size 0x4 offset 0x6 count 0x2 xfs_io-2946 [001] 253971.751415: xfs_iomap_prealloc_size: dev 253:32 ino 0x84 prealloc blocks 128 shift 0 m_writeio_blocks 16 xfs_io-2946 [001] 253971.751425: xfs_iomap_alloc: dev 253:32 ino 0x84 size 0x4 offset 0x6 count 131072 type invalid startoff 0x60 startblock -1 blockcount 0x90 That's the output I need for the complete test - you'll need to use a better recording mechanism that this (e.g. trace-cmd record, trace-cmd report) because it will generate a lot of events. Compress the two report files (they'll be large) and send them to me offlist. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
e 0x4 offset 0x4 count 0x2 xfs_io-2946 [001] 253971.751236: xfs_iomap_found: dev 253:32 ino 0x84 size 0x4 offset 0x4 count 131072 type invalid startoff 0x0 startblock 24 blockcount 0x60 xfs_io-2946 [001] 253971.751381: xfs_file_buffered_write: dev 253:32 ino 0x84 size 0x4 offset 0x6 count 0x2 xfs_io-2946 [001] 253971.751415: xfs_iomap_prealloc_size: dev 253:32 ino 0x84 prealloc blocks 128 shift 0 m_writeio_blocks 16 xfs_io-2946 [001] 253971.751425: xfs_iomap_alloc: dev 253:32 ino 0x84 size 0x4 offset 0x6 count 131072 type invalid startoff 0x60 startblock -1 blockcount 0x90 That's the output I need for the complete test - you'll need to use a better recording mechanism that this (e.g. trace-cmd record, trace-cmd report) because it will generate a lot of events. Compress the two report files (they'll be large) and send them to me offlist. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
lock we previously didn't spin on at all. We really need instruction level perf profiles to understand this - I don't have a machine with this many cpu cores available locally, so I'm not sure I'm going to be able to make any progress tracking it down in the short term. Maybe the lkp team has more in-depth cpu usage profiles they can share? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

2016-08-10 Thread Dave Chinner
ll. We really need instruction level perf profiles to understand this - I don't have a machine with this many cpu cores available locally, so I'm not sure I'm going to be able to make any progress tracking it down in the short term. Maybe the lkp team has more in-depth cpu usage profiles they can share? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

[GIT PULL] xfs: reverse mapping support for 4.8-rc1

2016-08-06 Thread Dave Chinner
ode 100644 fs/xfs/xfs_rmap_item.c create mode 100644 fs/xfs/xfs_rmap_item.h create mode 100644 fs/xfs/xfs_trans_rmap.c -- Dave Chinner da...@fromorbit.com

[GIT PULL] xfs: reverse mapping support for 4.8-rc1

2016-08-06 Thread Dave Chinner
ode 100644 fs/xfs/xfs_rmap_item.c create mode 100644 fs/xfs/xfs_rmap_item.h create mode 100644 fs/xfs/xfs_trans_rmap.c -- Dave Chinner da...@fromorbit.com

Re: [bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-05 Thread Dave Chinner
On Fri, Aug 05, 2016 at 09:59:35PM +1000, Dave Chinner wrote: > On Fri, Aug 05, 2016 at 11:54:17AM +0100, Mel Gorman wrote: > > On Fri, Aug 05, 2016 at 09:11:10AM +1000, Dave Chinner wrote: > > > > diff --git a/mm/page_alloc.c b/mm/page_alloc.c > > > > index

Re: [bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-05 Thread Dave Chinner
On Fri, Aug 05, 2016 at 09:59:35PM +1000, Dave Chinner wrote: > On Fri, Aug 05, 2016 at 11:54:17AM +0100, Mel Gorman wrote: > > On Fri, Aug 05, 2016 at 09:11:10AM +1000, Dave Chinner wrote: > > > > diff --git a/mm/page_alloc.c b/mm/page_alloc.c > > > > index

Re: [bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-05 Thread Dave Chinner
On Fri, Aug 05, 2016 at 11:54:17AM +0100, Mel Gorman wrote: > On Fri, Aug 05, 2016 at 09:11:10AM +1000, Dave Chinner wrote: > > > diff --git a/mm/page_alloc.c b/mm/page_alloc.c > > > index fb975cec3518..baa97da3687d 100644 > > > --- a/mm/page_alloc.c > > >

Re: [bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-05 Thread Dave Chinner
On Fri, Aug 05, 2016 at 11:54:17AM +0100, Mel Gorman wrote: > On Fri, Aug 05, 2016 at 09:11:10AM +1000, Dave Chinner wrote: > > > diff --git a/mm/page_alloc.c b/mm/page_alloc.c > > > index fb975cec3518..baa97da3687d 100644 > > > --- a/mm/page_alloc.c > > >

Re: [bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-04 Thread Dave Chinner
On Thu, Aug 04, 2016 at 01:34:58PM +0100, Mel Gorman wrote: > On Thu, Aug 04, 2016 at 01:24:09PM +0100, Mel Gorman wrote: > > On Thu, Aug 04, 2016 at 03:10:51PM +1000, Dave Chinner wrote: > > > Hi folks, > > > > > > I just noticed a whacky memory usage prof

Re: [bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-04 Thread Dave Chinner
On Thu, Aug 04, 2016 at 01:34:58PM +0100, Mel Gorman wrote: > On Thu, Aug 04, 2016 at 01:24:09PM +0100, Mel Gorman wrote: > > On Thu, Aug 04, 2016 at 03:10:51PM +1000, Dave Chinner wrote: > > > Hi folks, > > > > > > I just noticed a whacky memory usage prof

[bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-03 Thread Dave Chinner
and removed from teh page cache. According to the per-node counters, that is not happening and there gigabytes of invalidated pages still sitting on the active LRUs. Something is broken Cheers, Dave. -- Dave Chinner da...@fromorbit.com

[bug, 4.8] /proc/meminfo: counter values are very wrong

2016-08-03 Thread Dave Chinner
and removed from teh page cache. According to the per-node counters, that is not happening and there gigabytes of invalidated pages still sitting on the active LRUs. Something is broken Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 1/3] Add a new field to struct shrinker

2016-07-28 Thread Dave Chinner
On Thu, Jul 28, 2016 at 11:25:13AM +0100, Mel Gorman wrote: > On Thu, Jul 28, 2016 at 03:49:47PM +1000, Dave Chinner wrote: > > Seems you're all missing the obvious. > > > > Add a tracepoint for a shrinker callback that includes a "name" > > field, h

Re: [PATCH 1/3] Add a new field to struct shrinker

2016-07-28 Thread Dave Chinner
On Thu, Jul 28, 2016 at 11:25:13AM +0100, Mel Gorman wrote: > On Thu, Jul 28, 2016 at 03:49:47PM +1000, Dave Chinner wrote: > > Seems you're all missing the obvious. > > > > Add a tracepoint for a shrinker callback that includes a "name" > > field, h

Re: [PATCH 1/3] Add a new field to struct shrinker

2016-07-27 Thread Dave Chinner
cludes a "name" field, have the shrinker callback fill it out appropriately. e.g in the superblock shrinker: trace_shrinker_callback(shrinker, shrink_control, sb->s_type->name); And generic code that doesn't want to put a specific context name in there can simply call: trace_shrinker_callback(shrinker, shrink_control, __func__); And now you know exactly what shrinker is being run. No need to add names to any structures, it's call site defined so is flexible, and if you're not using tracepoints has no overhead. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 1/3] Add a new field to struct shrinker

2016-07-27 Thread Dave Chinner
ot; field, have the shrinker callback fill it out appropriately. e.g in the superblock shrinker: trace_shrinker_callback(shrinker, shrink_control, sb->s_type->name); And generic code that doesn't want to put a specific context name in there can simply call: trace_shrinker_callback(shrinker, shrink_control, __func__); And now you know exactly what shrinker is being run. No need to add names to any structures, it's call site defined so is flexible, and if you're not using tracepoints has no overhead. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

[GIT PULL] xfs: update for 4.8-rc1

2016-07-26 Thread Dave Chinner
rs xfs: convert list of extents to free into a regular list xfs: refactor btree maxlevels computation Dave Chinner (14): xfs: reduce lock hold times in buffer writeback Merge branch 'fs-4.8-iomap-infrastructure' into for-next Merge branch 'xfs-4.8-iomap-write' into fo

[GIT PULL] xfs: update for 4.8-rc1

2016-07-26 Thread Dave Chinner
rs xfs: convert list of extents to free into a regular list xfs: refactor btree maxlevels computation Dave Chinner (14): xfs: reduce lock hold times in buffer writeback Merge branch 'fs-4.8-iomap-infrastructure' into for-next Merge branch 'xfs-4.8-iomap-write' into fo

Re: [PATCH v3 1/4] lib/dlock-list: Distributed and lock-protected lists

2016-07-20 Thread Dave Chinner
ote. So it's really only a per-cpu structure for list addition Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH v3 1/4] lib/dlock-list: Distributed and lock-protected lists

2016-07-20 Thread Dave Chinner
ote. So it's really only a per-cpu structure for list addition Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: linux-next: manual merge of the xfs tree with Linus' tree

2016-07-20 Thread Dave Chinner
on't get the immediate attention of my mail filters, so I didn't see it immediately. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: linux-next: manual merge of the xfs tree with Linus' tree

2016-07-20 Thread Dave Chinner
on't get the immediate attention of my mail filters, so I didn't see it immediately. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [BUG] Slab corruption during XFS writeback under memory pressure

2016-07-19 Thread Dave Chinner
On Tue, Jul 19, 2016 at 02:22:47PM -0700, Calvin Owens wrote: > On 07/18/2016 07:05 PM, Calvin Owens wrote: > >On 07/17/2016 11:02 PM, Dave Chinner wrote: > >>On Sun, Jul 17, 2016 at 10:00:03AM +1000, Dave Chinner wrote: > >>>On Fri, Jul 15, 2016 at 05:18:

Re: [BUG] Slab corruption during XFS writeback under memory pressure

2016-07-19 Thread Dave Chinner
On Tue, Jul 19, 2016 at 02:22:47PM -0700, Calvin Owens wrote: > On 07/18/2016 07:05 PM, Calvin Owens wrote: > >On 07/17/2016 11:02 PM, Dave Chinner wrote: > >>On Sun, Jul 17, 2016 at 10:00:03AM +1000, Dave Chinner wrote: > >>>On Fri, Jul 15, 2016 at 05:18:

Re: [BUG] Slab corruption during XFS writeback under memory pressure

2016-07-18 Thread Dave Chinner
On Sun, Jul 17, 2016 at 10:00:03AM +1000, Dave Chinner wrote: > On Fri, Jul 15, 2016 at 05:18:02PM -0700, Calvin Owens wrote: > > Hello all, > > > > I've found a nasty source of slab corruption. Based on seeing similar > > symptoms > > on boxes at Facebook,

Re: [BUG] Slab corruption during XFS writeback under memory pressure

2016-07-18 Thread Dave Chinner
On Sun, Jul 17, 2016 at 10:00:03AM +1000, Dave Chinner wrote: > On Fri, Jul 15, 2016 at 05:18:02PM -0700, Calvin Owens wrote: > > Hello all, > > > > I've found a nasty source of slab corruption. Based on seeing similar > > symptoms > > on boxes at Facebook,

Re: [BUG] Slab corruption during XFS writeback under memory pressure

2016-07-16 Thread Dave Chinner
; if (fd == -1) { > perror("Can't open"); > return 1; > } > > if (!fork()) { > count = atol(argv[2]); > > while (1) { > for (i = 0; i < count; i++) > if (write(fd, crap, CHUNK) != CHUNK) > perror("Eh?"); > > fsync(fd); > ftruncate(fd, 0); > } H. Truncate is used, but only after fsync. If the truncate is removed, does the problem go away? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [BUG] Slab corruption during XFS writeback under memory pressure

2016-07-16 Thread Dave Chinner
; if (fd == -1) { > perror("Can't open"); > return 1; > } > > if (!fork()) { > count = atol(argv[2]); > > while (1) { > for (i = 0; i < count; i++) > if (write(fd, crap, CHUNK) != CHUNK) > perror("Eh?"); > > fsync(fd); > ftruncate(fd, 0); > } H. Truncate is used, but only after fsync. If the truncate is removed, does the problem go away? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

2016-07-11 Thread Dave Chinner
On Mon, Jul 11, 2016 at 10:02:24AM +0100, Mel Gorman wrote: > On Mon, Jul 11, 2016 at 10:47:57AM +1000, Dave Chinner wrote: > > > I had tested XFS with earlier releases and noticed no major problems > > > so later releases tested only one filesystem. Given the changes

Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

2016-07-11 Thread Dave Chinner
On Mon, Jul 11, 2016 at 10:02:24AM +0100, Mel Gorman wrote: > On Mon, Jul 11, 2016 at 10:47:57AM +1000, Dave Chinner wrote: > > > I had tested XFS with earlier releases and noticed no major problems > > > so later releases tested only one filesystem. Given the changes

Re: Hang due to nfs letting tasks freeze with locked inodes

2016-07-10 Thread Dave Chinner
On Fri, Jul 08, 2016 at 01:05:40PM +, Trond Myklebust wrote: > > On Jul 8, 2016, at 08:55, Trond Myklebust > > <tron...@primarydata.com> wrote: > >> On Jul 8, 2016, at 08:48, Seth Forshee > >> <seth.fors...@canonical.com> wrote: On Fri, Jul 08, 2016 at

Re: Hang due to nfs letting tasks freeze with locked inodes

2016-07-10 Thread Dave Chinner
On Fri, Jul 08, 2016 at 01:05:40PM +, Trond Myklebust wrote: > > On Jul 8, 2016, at 08:55, Trond Myklebust > > wrote: > >> On Jul 8, 2016, at 08:48, Seth Forshee > >> wrote: On Fri, Jul 08, 2016 at > >> 09:53:30AM +1000, Dave Chinner wrote: > &g

Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

2016-07-10 Thread Dave Chinner
On Fri, Jul 08, 2016 at 10:52:03AM +0100, Mel Gorman wrote: > On Fri, Jul 08, 2016 at 09:27:13AM +1000, Dave Chinner wrote: > > . > > > This series is not without its hazards. There are at least three areas > > > that I'm concerned with even though I could

Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

2016-07-10 Thread Dave Chinner
On Fri, Jul 08, 2016 at 10:52:03AM +0100, Mel Gorman wrote: > On Fri, Jul 08, 2016 at 09:27:13AM +1000, Dave Chinner wrote: > > . > > > This series is not without its hazards. There are at least three areas > > > that I'm concerned with even though I could

Re: Hang due to nfs letting tasks freeze with locked inodes

2016-07-07 Thread Dave Chinner
sys_sync() isn't sufficient to quiesce a filesystem's operations. But I'm used to being ignored on this topic (for almost 10 years, now!). Indeed, it's been made clear in the past that I know absolutely nothing about what is needed to be done to safely suspend filesystem operations... :/ Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: Hang due to nfs letting tasks freeze with locked inodes

2016-07-07 Thread Dave Chinner
sys_sync() isn't sufficient to quiesce a filesystem's operations. But I'm used to being ignored on this topic (for almost 10 years, now!). Indeed, it's been made clear in the past that I know absolutely nothing about what is needed to be done to safely suspend filesystem operations... :/ Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

2016-07-07 Thread Dave Chinner
sts you ran on ext4. It might also be worth running some highly concurrent inode cache benchmarks (e.g. the 50-million inode, 16-way concurrent fsmark tests) to see what impact heavy slab cache pressure has on shrinker behaviour and system balance... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

2016-07-07 Thread Dave Chinner
sts you ran on ext4. It might also be worth running some highly concurrent inode cache benchmarks (e.g. the 50-million inode, 16-way concurrent fsmark tests) to see what impact heavy slab cache pressure has on shrinker behaviour and system balance... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [Cluster-devel] [PATCH 2/2] GFS2: Add a gfs2-specific prune_icache_sb

2016-06-28 Thread Dave Chinner
On Tue, Jun 28, 2016 at 10:13:32AM +0100, Steven Whitehouse wrote: > Hi, > > On 28/06/16 03:08, Dave Chinner wrote: > >On Fri, Jun 24, 2016 at 02:50:11PM -0500, Bob Peterson wrote: > >>This patch adds a new prune_icache_sb function for the VFS slab > >>shrinker

Re: [Cluster-devel] [PATCH 2/2] GFS2: Add a gfs2-specific prune_icache_sb

2016-06-28 Thread Dave Chinner
On Tue, Jun 28, 2016 at 10:13:32AM +0100, Steven Whitehouse wrote: > Hi, > > On 28/06/16 03:08, Dave Chinner wrote: > >On Fri, Jun 24, 2016 at 02:50:11PM -0500, Bob Peterson wrote: > >>This patch adds a new prune_icache_sb function for the VFS slab > >>shrinker

Re: [PATCH 2/2] GFS2: Add a gfs2-specific prune_icache_sb

2016-06-27 Thread Dave Chinner
then move the parts of inode *freeing* that cause problems to a different context via the ->evict/destroy callouts and trigger that external context processing on demand. That external context can just do bulk "if it is on the list then free it" processing, because the reclaim policy has already been executed to place that inode on the reclaim list. This is essentially what XFS does, but it also uses the ->nr_cached_objects/->free_cached_objects() callouts in the superblock shrinker to provide the reclaim rate feedback mechanism required to throttle incoming memory allocations. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 2/2] GFS2: Add a gfs2-specific prune_icache_sb

2016-06-27 Thread Dave Chinner
*freeing* that cause problems to a different context via the ->evict/destroy callouts and trigger that external context processing on demand. That external context can just do bulk "if it is on the list then free it" processing, because the reclaim policy has already been executed to place that inode on the reclaim list. This is essentially what XFS does, but it also uses the ->nr_cached_objects/->free_cached_objects() callouts in the superblock shrinker to provide the reclaim rate feedback mechanism required to throttle incoming memory allocations. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 1/2] vfs: Add hooks for filesystem-specific prune_icache_sb

2016-06-27 Thread Dave Chinner
n of the inode but do not destroy/free it - you simply queue it to an internal list and then do the cleanup/freeing in your own time? i.e. why do you need a special callout just to defer freeing to another thread when we already have hooks than enable you to do this? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [PATCH 1/2] vfs: Add hooks for filesystem-specific prune_icache_sb

2016-06-27 Thread Dave Chinner
ot destroy/free it - you simply queue it to an internal list and then do the cleanup/freeing in your own time? i.e. why do you need a special callout just to defer freeing to another thread when we already have hooks than enable you to do this? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [RFC PATCH 2/2] xfs: map KM_MAYFAIL to __GFP_RETRY_HARD

2016-06-15 Thread Dave Chinner
ailure, so retry on failure is not required." To then map KM_MAYFAIL to a flag that implies the allocation will internally retry to try exceptionally hard to prevent failure seems wrong. IOWs, KM_MAYFAIL means XFS is just using for normal allocator behaviour here, so I'm not sure what problem

Re: [RFC PATCH 2/2] xfs: map KM_MAYFAIL to __GFP_RETRY_HARD

2016-06-15 Thread Dave Chinner
ed." To then map KM_MAYFAIL to a flag that implies the allocation will internally retry to try exceptionally hard to prevent failure seems wrong. IOWs, KM_MAYFAIL means XFS is just using for normal allocator behaviour here, so I'm not sure what problem this change is actually solving and it's not clear from the description Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [RFC PATCH-tip 6/6] xfs: Enable reader optimistic spinning for DAX inodes

2016-06-14 Thread Dave Chinner
*extremely* paranoid when it comes to changes to core locking like this. Performance is secondary to correctness, and we need much more than just a few benchmarks to verify there aren't locking bugs being introduced Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [RFC PATCH-tip 6/6] xfs: Enable reader optimistic spinning for DAX inodes

2016-06-14 Thread Dave Chinner
comes to changes to core locking like this. Performance is secondary to correctness, and we need much more than just a few benchmarks to verify there aren't locking bugs being introduced Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-06-02 Thread Dave Chinner
On Thu, Jun 02, 2016 at 02:44:30PM +0200, Holger Hoffstätte wrote: > On 06/02/16 14:13, Stefan Priebe - Profihost AG wrote: > > > > Am 31.05.2016 um 09:31 schrieb Dave Chinner: > >> On Tue, May 31, 2016 at 08:11:42AM +0200, Stefan Priebe - Profihost AG > >&g

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-06-02 Thread Dave Chinner
On Thu, Jun 02, 2016 at 02:44:30PM +0200, Holger Hoffstätte wrote: > On 06/02/16 14:13, Stefan Priebe - Profihost AG wrote: > > > > Am 31.05.2016 um 09:31 schrieb Dave Chinner: > >> On Tue, May 31, 2016 at 08:11:42AM +0200, Stefan Priebe - Profihost AG > >&g

Re: Internal error xfs_trans_cancel

2016-06-02 Thread Dave Chinner
ith the same steps. Hmmm, Ok. I've been running the lockperf test and kernel builds all day on a filesystem that is identical in shape and size to yours (i.e. xfs_info output is the same) but I haven't reproduced it yet. Is it possible to get a metadump image of your filesystem to see if I can reproduce it on that? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: Internal error xfs_trans_cancel

2016-06-02 Thread Dave Chinner
ith the same steps. Hmmm, Ok. I've been running the lockperf test and kernel builds all day on a filesystem that is identical in shape and size to yours (i.e. xfs_info output is the same) but I haven't reproduced it yet. Is it possible to get a metadump image of your filesystem to see if I can reproduce it on that? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: Internal error xfs_trans_cancel

2016-06-01 Thread Dave Chinner
597's new affinity list: 0,4,8,12 sh: 1: cannot create /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor: Directory nonexistent posix01 -n 8 -l 100 posix02 -n 8 -l 100 posix03 -n 8 -i 100 $ So, I've just removed those tests from your script. I'll see if I have any luck with reproducing the problem now. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: Internal error xfs_trans_cancel

2016-06-01 Thread Dave Chinner
597's new affinity list: 0,4,8,12 sh: 1: cannot create /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor: Directory nonexistent posix01 -n 8 -l 100 posix02 -n 8 -l 100 posix03 -n 8 -i 100 $ So, I've just removed those tests from your script. I'll see if I have any luck with reproducing the problem now. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: Internal error xfs_trans_cancel

2016-06-01 Thread Dave Chinner
ormation_should_I_include_when_reporting_a_problem.3F You didn't run out of space or something unusual like that? Does 'xfs_repair -n ' report any errors? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: Internal error xfs_trans_cancel

2016-06-01 Thread Dave Chinner
ormation_should_I_include_when_reporting_a_problem.3F You didn't run out of space or something unusual like that? Does 'xfs_repair -n ' report any errors? Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-31 Thread Dave Chinner
e appears to be handling the dirty page that is being passed to it correctly. We'll work out what needs to be done to get rid of the warning for this case, wether it be a mm/ change or an XFS change. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-31 Thread Dave Chinner
e appears to be handling the dirty page that is being passed to it correctly. We'll work out what needs to be done to get rid of the warning for this case, wether it be a mm/ change or an XFS change. Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-31 Thread Dave Chinner
On Tue, May 31, 2016 at 12:59:04PM +0900, Minchan Kim wrote: > On Tue, May 31, 2016 at 12:55:09PM +1000, Dave Chinner wrote: > > On Tue, May 31, 2016 at 10:07:24AM +0900, Minchan Kim wrote: > > > On Tue, May 31, 2016 at 08:36:57AM +1000, Dave Chinner wrote: > > > > B

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-31 Thread Dave Chinner
On Tue, May 31, 2016 at 12:59:04PM +0900, Minchan Kim wrote: > On Tue, May 31, 2016 at 12:55:09PM +1000, Dave Chinner wrote: > > On Tue, May 31, 2016 at 10:07:24AM +0900, Minchan Kim wrote: > > > On Tue, May 31, 2016 at 08:36:57AM +1000, Dave Chinner wrote: > > > > B

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-30 Thread Dave Chinner
On Tue, May 31, 2016 at 10:07:24AM +0900, Minchan Kim wrote: > On Tue, May 31, 2016 at 08:36:57AM +1000, Dave Chinner wrote: > > [adding lkml and linux-mm to the cc list] > > > > On Mon, May 30, 2016 at 09:23:48AM +0200, Stefan Priebe - Profihost AG > > wrote: >

Re: shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-30 Thread Dave Chinner
On Tue, May 31, 2016 at 10:07:24AM +0900, Minchan Kim wrote: > On Tue, May 31, 2016 at 08:36:57AM +1000, Dave Chinner wrote: > > [adding lkml and linux-mm to the cc list] > > > > On Mon, May 30, 2016 at 09:23:48AM +0200, Stefan Priebe - Profihost AG > > wrote: >

shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-30 Thread Dave Chinner
and memory reclaim. It might be worth trying as a workaround for now. MM-folk - is this analysis correct? If so, why is shrink_active_list() calling try_to_release_page() on dirty pages? Is this just an oversight or is there some problem that this is trying to work around? It seems trivial to fix to me (add a !PageDirty check), but I don't know why the check is there in the first place... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

shrink_active_list/try_to_release_page bug? (was Re: xfs trace in 4.4.2 / also in 4.3.3 WARNING fs/xfs/xfs_aops.c:1232 xfs_vm_releasepage)

2016-05-30 Thread Dave Chinner
and memory reclaim. It might be worth trying as a workaround for now. MM-folk - is this analysis correct? If so, why is shrink_active_list() calling try_to_release_page() on dirty pages? Is this just an oversight or is there some problem that this is trying to work around? It seems trivial to fix to me (add a !PageDirty check), but I don't know why the check is there in the first place... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [GIT PULL] xfs: updates for 4.7-rc1

2016-05-29 Thread Dave Chinner
On Thu, May 26, 2016 at 07:05:11PM -0700, Linus Torvalds wrote: > On Thu, May 26, 2016 at 5:13 PM, Dave Chinner <da...@fromorbit.com> wrote: > > On Thu, May 26, 2016 at 10:19:13AM -0700, Linus Torvalds wrote: > >> > >> i'm ok with the late branches, it's not

Re: [GIT PULL] xfs: updates for 4.7-rc1

2016-05-29 Thread Dave Chinner
On Thu, May 26, 2016 at 07:05:11PM -0700, Linus Torvalds wrote: > On Thu, May 26, 2016 at 5:13 PM, Dave Chinner wrote: > > On Thu, May 26, 2016 at 10:19:13AM -0700, Linus Torvalds wrote: > >> > >> i'm ok with the late branches, it's not like xfs has been a problem sp

Re: [GIT PULL] xfs: updates for 4.7-rc1

2016-05-26 Thread Dave Chinner
On Thu, May 26, 2016 at 10:19:13AM -0700, Linus Torvalds wrote: > On Wed, May 25, 2016 at 11:13 PM, Dave Chinner <da...@fromorbit.com> wrote: > > > > Just yell if this is not OK and I'll drop those branches for this > > merge and resend the pull request > > i'

Re: [GIT PULL] xfs: updates for 4.7-rc1

2016-05-26 Thread Dave Chinner
On Thu, May 26, 2016 at 10:19:13AM -0700, Linus Torvalds wrote: > On Wed, May 25, 2016 at 11:13 PM, Dave Chinner wrote: > > > > Just yell if this is not OK and I'll drop those branches for this > > merge and resend the pull request > > i'm ok with the late bran

[GIT PULL] xfs: updates for 4.7-rc1

2016-05-26 Thread Dave Chinner
warning in xfs_finish_page_writeback for non-debug builds Dave Chinner (20): xfs: Don't wrap growfs AGFL indexes xfs: build bios directly in xfs_add_to_ioend xfs: don't release bios on completion immediately xfs: remove xfs_fs_evict_inode() xfs: xfs_iflush_cluster fa

[GIT PULL] xfs: updates for 4.7-rc1

2016-05-26 Thread Dave Chinner
warning in xfs_finish_page_writeback for non-debug builds Dave Chinner (20): xfs: Don't wrap growfs AGFL indexes xfs: build bios directly in xfs_add_to_ioend xfs: don't release bios on completion immediately xfs: remove xfs_fs_evict_inode() xfs: xfs_iflush_cluster fa

Re: [GIT PULL] y2038 changes for vfs

2016-05-25 Thread Dave Chinner
et *exactly* like Linus us now suggesting, I walked away and haven't looked at your patches since. Is it any wonder that no other filesystem maintainer has bothered to waste their time on this since? Linus - I'd suggest these VFS timestamp patches need to go through Al's VFS tree. That way we don't get unreviewed VFS infrastructure changes going into your tree via a door that nobody was paying attention to... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

Re: [GIT PULL] y2038 changes for vfs

2016-05-25 Thread Dave Chinner
e Linus us now suggesting, I walked away and haven't looked at your patches since. Is it any wonder that no other filesystem maintainer has bothered to waste their time on this since? Linus - I'd suggest these VFS timestamp patches need to go through Al's VFS tree. That way we don't get unreviewed VFS infrastructure changes going into your tree via a door that nobody was paying attention to... Cheers, Dave. -- Dave Chinner da...@fromorbit.com

<    8   9   10   11   12   13   14   15   16   17   >