Hi JFS Developers,
I was unable to find much information about this issue so please shed some 
light on what is going on.
We are running Debian Wheezy 64bit, kernel: 3.0.0-1-amd64 #1 SMP, JFS: 1.1.15
We have a JFS filesystem on an LVM volume which is on a HW RAID-6 Array.  The 
size of the LVM and JFS filesystem is 1.5TB.  JFS filesystem was created with 
default options.  We have a process that created ~100-200 million small files 
(under 4kb) on this volume.  From what I read 200 million files should not be 
an issue for JFS, correct?  A lot of these files ended up in the same directory 
and running 'ls > file' on this directory took ~7hrs.  We did have a couple of 
unplanned reboots on this server so I expected some metadata corruption. 
fsck completed with no erros but running jfs_logdump -a <volume> produced this 
output:jfs_logdump version 1.1.15, 04-Mar-2011Device Name: 
/dev/blah/blah1LOGREDO:  The Journal Log has wrapped. [logredo.c:1339]LOGREDO:  
 logRead: Log wrapped over itself (lognumread = (d) 8191). [log_read.c:377]log 
read failed 0x14d3f98JFS_LOGDUMP: The current JFS log has been dumped into 
./jfslog.dmp
Is this a "fatal" error?  or is this normal JFS log operation considering the 
amount of files that were created?
I read that the maximum log size for JFS is 128MB and since we used default 
options (0.04% of 1.5TB is well over 128MB) our log should be the maximum 
already.  This implies that I cannot increase the log-size.
So please shed some light on what the log-wrap condition means.  I realize that 
putting this many files into one directory is not ideal but is it just a matter 
of performance or am I loosing/corrupting data?
let me know if I need to provide additional information.
thank you,
-stan                                     
------------------------------------------------------------------------------
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2dcopy1
_______________________________________________
Jfs-discussion mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/jfs-discussion

Reply via email to