Yes, please file a bugzilla and upload stat_sysdir.sh. url to log a bugzilla is http://oss.oracle.com/bugzilla/enter_bug.cgi?product=OCFS2

Juan Pablo Marco Cardona wrote:
Hi,
you are right, i can create a 5GB file but no more.
But the trouble isn't in the file size, it is in the total free size os the partition.

For example if a i create 2 file of 2.5 GB, i can't create a third file of 2.5 GB.

If after creation, i remove and create again the file happens the same.

Really, i don't know what is the problem!

Tomorrow, like Sunil said me, i will open a bug report in Bugzilla attaching the stat_sysdir.sh output.

What is the url bugzilla to open the bug report about this problem?

Thanks in advanced

Regards,
Pablo

P.D:
Now just i cant write ~ 2.8 GB in the same partition:

df -kh | grep mail
300G  239G   62G  80% /home/mail

dd if=/dev/zero of=./file1 bs=100M count=25
25+0 records in
25+0 records out
2621440000 bytes (2,6 GB) copied, 129,77 seconds, 20,2 MB/s
dd if=/dev/zero of=./file2 bs=100M count=25 dd: escribiendo «./file2»: No queda espacio en el dispositivo (Spanish message about the lack of free space)
3+0 records in
2+0 records out
251658240 bytes (252 MB) copied, 3,52704 seconds, 71,4 MB/s



Srinivas Eeda escribió:
I don't see any problem with your filesystem configuration. df reports 63G free space, so you should be able to create files of that size(approximately). But you are able to create one file of size 5gb and no more? Or you are able to create another 5GB file?. Once you get ENOSPC what happens, can't create anymore? What happens if you delete the 5GB file, you can create that again?

4k cluster size and 1k block size is good. But if your file sizes are bigger N*4K you can set the clustersize to that so you can benefit more.

Juan Pablo Marco Cardona wrote:
Hi Srinivas,
the trouble happens creating files like this:

dd if=/dev/zero of=./file bs=100M count=60

When the file arrives to 5GB the filesystem outputs a error about no more free space!

Theorically there is about 63 GB of free space:

df -kh | grep mail
300G  238G   63G  80% /home/mail

This trouble happens creating small files, too

The inodes seems not to be exhausted:

df -i | grep mail
78642183 62365063 16277120   80% /home/mail

I think, there are so few orphaned inodes and maybe it doesn't affect:

echo "ls //orphan_dir:0000" | debugfs.ocfs2 /dev/sda1

debugfs.ocfs2 1.4.3
debugfs:        16              16   1    2  .
        10              16   2    2  ..
        79813023        28   16   1  0000000004c1d99f
        28472617        28   16   1  0000000001b27529
        8438318         28   16   1  000000000080c22e
        80973610        28   16   1  0000000004d38f2a
        213406992       28   16   1  000000000cb85510
        13234609        28   16   1  0000000000c9f1b1
        228704523       28   16   1  000000000da1c10b
        225968869       28   16   1  000000000d7802e5
        256692752       28   16   1  000000000f4cd210
        103224605       28   16   1  000000000627151d
        83675914        28   16   1  0000000004fccb0a
        225968588       28   16   1  000000000d7801cc
        278103558       28   16   1  0000000010938606
        256692760       28   16   1  000000000f4cd218
        13235439        28   16   1  0000000000c9f4ef
        8861985         28   16   1  0000000000873921
        228997031       28   16   1  000000000da637a7
        111786205       28   16   1  0000000006a9b8dd
        24850261        28   16   1  00000000017b2f55
        29889095        28   16   1  0000000001c81247
        311022924       28   16   1  000000001289d54c
        13235309        28   16   1  0000000000c9f46d
        129665645       28   16   1  0000000007ba8a6d
        79605831        28   16   1  0000000004beb047
        28104589        28   16   1  0000000001acd78d
        294769884       28   16   1  000000001191d4dc
        253519012       28   16   1  000000000f1c64a4
        80973232        28   16   1  0000000004d38db0
        13234376        28   16   1  0000000000c9f0c8
        312527073       28   16   1  0000000012a0c8e1
        25863407        28   16   1  00000000018aa4ef
        305612210       28   16   1  00000000123745b2
        226494741       28   16   1  000000000d800915
        228705439       28   16   1  000000000da1c49f
        79604433        40   16   1  0000000004beaad1
        80973605        28   16   1  0000000004d38f25
        226494890       28   16   1  000000000d8009aa
        80973659        28   16   1  0000000004d38f5b
        13236194        56   16   1  0000000000c9f7e2
        80973320        56   16   1  0000000004d38e08
        46226861        28   16   1  0000000002c15dad
        228995934       28   16   1  000000000da6335e
        294770302       28   16   1  000000001191d67e
        256692906       28   16   1  000000000f4cd2aa
        255321636       56   16   1  000000000f37e624
        80974273        28   16   1  0000000004d391c1
        209062548       56   16   1  000000000c760a94
        46227042        168  16   1  0000000002c15e62
        108197615       28   16   1  000000000672f6ef
        13236185        56   16   1  0000000000c9f7d9
        294768980       112  16   1  000000001191d154
        294768983       212  16   1  000000001191d157


I will show you some info about the partition:

Block size:

tunefs.ocfs2 -q -Q "BS=%5B\n" /dev/sda1
BS= 1024

Cluster size:

tunefs.ocfs2 -q -Q "CS=%5T\n" /dev/sda1
CS= 4096

Number Cluster Nodes:

tunefs.ocfs2 -q -Q "CN=%5N\n" /dev/sda1
CN=    5

I think is better to give you more information about this partition.

In this 300 GB ocfs2 partition are the maildir's of about 3K users with a lot of small files ( ~ 62 M used inodes). In the near future we are planning to mount a 2-mail-cluster system, because this volume is in a SAN.

Perhaps the cluster and node size are to much big? Maybe the trouble is fragmentation?
What are the best ocfs2 options to format this mail volume?

Thanks in advanced.

Regards,
Pablo




Srinivas Eeda escribió:
Do you run into ENOSPC creating new files or extending existing files? What is the cluster size? Don't think this may be the issue, but any files under orphan directory? run (echo "ls //orphan_dir:000X" | debugfs.ocfs2 <device>) to check if there are any.


Juan Pablo Marco Cardona wrote:
Hi Sunil,
we still have problems with the ocfs2 mail partition. The old problem with the free space still continues.
Theorically we hace 62 GB free space, but we only can use about 5 GB!

df -kh | grep mail
300G  238G   63G  80% /home/mail

Also, yesterday we have a power outage that force a fsck.ocfs2 of the partition.

We have upgraded to ocfs2 1.4.4 and ocfs2 1.4.3 tools and the kernel version is the same:

rpm -qa | grep ocfs2
ocfs2-2.6.18-128.2.1.el5-1.4.4-1.el5
ocfs2-tools-1.4.3-1.el5
ocfs2console-1.4.3-1.el5

modinfo ocfs2 | grep ^version
version:        1.4.4

uname -r
2.6.18-128.2.1.el5

What we can make? Maybe make a fsck.ocfs2 or mkfs.ocfs2 with the new tools version (1.4.3) ??

Thanks in advance.

Regards,
Pablo







--

------------------------------------------------------------------------

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users


--

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users

Reply via email to