Re: [Gluster-devel] regarding inode_link/unlink

2014-07-05 Thread Raghavendra Gowdappa


- Original Message -
> From: "Pranith Kumar Karampuri" 
> To: "Raghavendra Gowdappa" 
> Cc: "Gluster Devel" , "Anand Avati" 
> , "Brian Foster"
> , "Raghavendra Bhat" 
> Sent: Friday, July 4, 2014 5:39:03 PM
> Subject: Re: regarding inode_link/unlink
> 
> 
> On 07/04/2014 04:28 PM, Raghavendra Gowdappa wrote:
> >
> > - Original Message -
> >> From: "Pranith Kumar Karampuri" 
> >> To: "Gluster Devel" , "Anand Avati"
> >> , "Brian Foster"
> >> , "Raghavendra Gowdappa" ,
> >> "Raghavendra Bhat" 
> >> Sent: Friday, July 4, 2014 3:44:29 PM
> >> Subject: regarding inode_link/unlink
> >>
> >> hi,
> >>I have a doubt about when a particular dentry_unset thus
> >> inode_unref on parent dir happens on fuse-bridge in gluster.
> >> When a file is looked up for the first time fuse_entry_cbk does
> >> 'inode_link' with parent-gfid/bname. Whenever an unlink/rmdir/(lookup
> >> gives ENOENT) happens then corresponding inode unlink happens. The
> >> question is, will the present set of operations lead to leaks:
> >> 1) Mount 'M0' creates a file 'a'
> >> 2) Mount 'M1' of same volume deletes file 'a'
> >>
> >> M0 never touches 'a' anymore. When will inode_unlink happen for such
> >> cases? Will it lead to memory leaks?
> > Kernel will eventually send forget (a) on M0 and that will cleanup the
> > dentries and inode. Its equivalent to a file being looked up and never
> > used again (deleting doesn't matter in this case).
> Do you know the trigger points for that? When I do 'touch a' on the
> mount point and leave the system like that, forget is not coming.
> If I do unlink on the file then forget is coming.

I am not very familiar with how kernel manages its inodes. However, as Avati 
has mentioned in another mail, you can force kernel to send forgets by 
invalidating the inode. I think he has given enough details in another mail.

> 
> Pranith
> >
> >
> >> Pranith
> >>
> 
> 
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] Problem with smoke, regression ordering

2014-07-05 Thread Pranith Kumar Karampuri

hi Justin,
 If the regression results complete before the smoke test then 
'green-tick-mark' is over-written and people don't realize that the 
regression succeeded by a simple glance at the list of patches. Can we 
do anything about it?


Pranith
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] Problem with smoke, regression ordering

2014-07-05 Thread Niels de Vos
On Sat, Jul 05, 2014 at 09:53:23PM +0530, Pranith Kumar Karampuri wrote:
> hi Justin,
>  If the regression results complete before the smoke test then
> 'green-tick-mark' is over-written and people don't realize that the
> regression succeeded by a simple glance at the list of patches. Can
> we do anything about it?

This would likely be solved if we delay starting the regression testing 
after a positive review was performed.

An other option is to enable more hosts to run the smoke tests. These 
are currently only run on the main system and that is also used for 
other jobs (some manual regression tests, release jobs, ...).

Do others have any other ideas or suggestions?

Niels
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


[Gluster-devel] FS Sanity daily results.

2014-07-05 Thread Benjamin Turner
Hi all.  I have been running FS sanity on daily builds(glusterfs mounts
only at this point) for a few days for a few days and I have been hitting a
couple of problems:

 final pass/fail report =
   Test Date: Sat Jul  5 01:53:00 EDT 2014
   Total : [44]
   Passed: [41]
   Failed: [3]
   Abort : [0]
   Crash : [0]
-
   [   PASS   ]  FS Sanity Setup
   [   PASS   ]  Running tests.
   [   PASS   ]  FS SANITY TEST - arequal
   [   PASS   ]  FS SANITY LOG SCAN - arequal
   [   PASS   ]  FS SANITY LOG SCAN - bonnie
   [   PASS   ]  FS SANITY TEST - glusterfs_build
   [   PASS   ]  FS SANITY LOG SCAN - glusterfs_build
   [   PASS   ]  FS SANITY TEST - compile_kernel
   [   PASS   ]  FS SANITY LOG SCAN - compile_kernel
   [   PASS   ]  FS SANITY TEST - dbench
   [   PASS   ]  FS SANITY LOG SCAN - dbench
   [   PASS   ]  FS SANITY TEST - dd
   [   PASS   ]  FS SANITY LOG SCAN - dd
   [   PASS   ]  FS SANITY TEST - ffsb
   [   PASS   ]  FS SANITY LOG SCAN - ffsb
   [   PASS   ]  FS SANITY TEST - fileop
   [   PASS   ]  FS SANITY LOG SCAN - fileop
   [   PASS   ]  FS SANITY TEST - fsx
   [   PASS   ]  FS SANITY LOG SCAN - fsx
   [   PASS   ]  FS SANITY LOG SCAN - fs_mark
   [   PASS   ]  FS SANITY TEST - iozone
   [   PASS   ]  FS SANITY LOG SCAN - iozone
   [   PASS   ]  FS SANITY TEST - locks
   [   PASS   ]  FS SANITY LOG SCAN - locks
   [   PASS   ]  FS SANITY TEST - ltp
   [   PASS   ]  FS SANITY LOG SCAN - ltp
   [   PASS   ]  FS SANITY TEST - multiple_files
   [   PASS   ]  FS SANITY LOG SCAN - multiple_files
   [   PASS   ]  FS SANITY TEST - posix_compliance
   [   PASS   ]  FS SANITY LOG SCAN - posix_compliance
   [   PASS   ]  FS SANITY TEST - postmark
   [   PASS   ]  FS SANITY LOG SCAN - postmark
   [   PASS   ]  FS SANITY TEST - read_large
   [   PASS   ]  FS SANITY LOG SCAN - read_large
   [   PASS   ]  FS SANITY TEST - rpc
   [   PASS   ]  FS SANITY LOG SCAN - rpc
   [   PASS   ]  FS SANITY TEST - syscallbench
   [   PASS   ]  FS SANITY LOG SCAN - syscallbench
   [   PASS   ]  FS SANITY TEST - tiobench
   [   PASS   ]  FS SANITY LOG SCAN - tiobench
   [   PASS   ]  FS Sanity Cleanup

   [   FAIL   ]  FS SANITY TEST - bonnie
   [   FAIL   ]  FS SANITY TEST - fs_mark
   [   FAIL   ]
/rhs-tests/beaker/rhs/auto-tests/components/sanity/fs-sanity-tests-v2


Bonnie++ is just very slow(running for 10+ hours on 1 16 GB file) and
FS mark has been failing.  The bonnie slowness is in re read, here is
the best explanation I can find on it:

https://blogs.oracle.com/roch/entry/decoding_bonnie

*Rewriting...done*

This gets a little interesting. It actually reads 8K, lseek back to
the start of the block, overwrites the 8K with new data and loops.
(see article for more.).

On FS mark I am seeing:

#  fs_mark  -d  .  -D  4  -t  4  -S  5
#   Version 3.3, 4 thread(s) starting at Sat Jul  5 00:54:00 2014
#   Sync method: POST: Reopen and fsync() each file in order after main
write loop.
#   Directories:  Time based hash between directories across 4
subdirectories with 180 seconds per subdirectory.
#   File names: 40 bytes long, (16 initial bytes of time stamp with 24
random bytes at end of name)
#   Files info: size 51200 bytes, written with an IO size of 16384 bytes 
per write
#   App overhead is time in microseconds spent in the test not doing
file writing related system calls.

FSUse%Count SizeFiles/sec App Overhead
Error in unlink of ./00/53b784e8SKZ0QS9BO7O2EG1DIFQLRDYY : No
such file or directory
fopen failed to open: fs_log.txt.26676
fs-mark pass # 5 failed

I am working on reporting so look for a daily status report email from
my jenkins server soon.  How do we want to handle failures like this
moving forward?  Should I just open a BZ after I triage?  Do you guys
do a new BZ for every failure in the normal regressions tests?


-b
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel