Hi Srinivas,
the trouble happens creating files like this:

dd if=/dev/zero of=./file bs=100M count=60

When the file arrives to 5GB the filesystem outputs a error about no more free space!

Theorically there is about 63 GB of free space:

df -kh | grep mail
300G  238G   63G  80% /home/mail

This trouble happens creating small files, too

The inodes seems not to be exhausted:

df -i | grep mail
78642183 62365063 16277120   80% /home/mail

I think, there are so few orphaned inodes and maybe it doesn't affect:

echo "ls //orphan_dir:0000" | debugfs.ocfs2 /dev/sda1

debugfs.ocfs2 1.4.3
debugfs:        16              16   1    2  .
       10              16   2    2  ..
       79813023        28   16   1  0000000004c1d99f
       28472617        28   16   1  0000000001b27529
       8438318         28   16   1  000000000080c22e
       80973610        28   16   1  0000000004d38f2a
       213406992       28   16   1  000000000cb85510
       13234609        28   16   1  0000000000c9f1b1
       228704523       28   16   1  000000000da1c10b
       225968869       28   16   1  000000000d7802e5
       256692752       28   16   1  000000000f4cd210
       103224605       28   16   1  000000000627151d
       83675914        28   16   1  0000000004fccb0a
       225968588       28   16   1  000000000d7801cc
       278103558       28   16   1  0000000010938606
       256692760       28   16   1  000000000f4cd218
       13235439        28   16   1  0000000000c9f4ef
       8861985         28   16   1  0000000000873921
       228997031       28   16   1  000000000da637a7
       111786205       28   16   1  0000000006a9b8dd
       24850261        28   16   1  00000000017b2f55
       29889095        28   16   1  0000000001c81247
       311022924       28   16   1  000000001289d54c
       13235309        28   16   1  0000000000c9f46d
       129665645       28   16   1  0000000007ba8a6d
       79605831        28   16   1  0000000004beb047
       28104589        28   16   1  0000000001acd78d
       294769884       28   16   1  000000001191d4dc
       253519012       28   16   1  000000000f1c64a4
       80973232        28   16   1  0000000004d38db0
       13234376        28   16   1  0000000000c9f0c8
       312527073       28   16   1  0000000012a0c8e1
       25863407        28   16   1  00000000018aa4ef
       305612210       28   16   1  00000000123745b2
       226494741       28   16   1  000000000d800915
       228705439       28   16   1  000000000da1c49f
       79604433        40   16   1  0000000004beaad1
       80973605        28   16   1  0000000004d38f25
       226494890       28   16   1  000000000d8009aa
       80973659        28   16   1  0000000004d38f5b
       13236194        56   16   1  0000000000c9f7e2
       80973320        56   16   1  0000000004d38e08
       46226861        28   16   1  0000000002c15dad
       228995934       28   16   1  000000000da6335e
       294770302       28   16   1  000000001191d67e
       256692906       28   16   1  000000000f4cd2aa
       255321636       56   16   1  000000000f37e624
       80974273        28   16   1  0000000004d391c1
       209062548       56   16   1  000000000c760a94
       46227042        168  16   1  0000000002c15e62
       108197615       28   16   1  000000000672f6ef
       13236185        56   16   1  0000000000c9f7d9
       294768980       112  16   1  000000001191d154
       294768983       212  16   1  000000001191d157


I will show you some info about the partition:

Block size:

tunefs.ocfs2 -q -Q "BS=%5B\n" /dev/sda1
BS= 1024

Cluster size:

tunefs.ocfs2 -q -Q "CS=%5T\n" /dev/sda1
CS= 4096

Number Cluster Nodes:

tunefs.ocfs2 -q -Q "CN=%5N\n" /dev/sda1
CN=    5

I think is better to give you more information about this partition.

In this 300 GB ocfs2 partition are the maildir's of about 3K users with a lot of small files ( ~ 62 M used inodes). In the near future we are planning to mount a 2-mail-cluster system, because this volume is in a SAN.

Perhaps the cluster and node size are to much big? Maybe the trouble is fragmentation?
What are the best ocfs2 options to format this mail volume?

Thanks in advanced.

Regards,
Pablo




Srinivas Eeda escribió:
Do you run into ENOSPC creating new files or extending existing files? What is the cluster size? Don't think this may be the issue, but any files under orphan directory? run (echo "ls //orphan_dir:000X" | debugfs.ocfs2 <device>) to check if there are any.


Juan Pablo Marco Cardona wrote:
Hi Sunil,
we still have problems with the ocfs2 mail partition. The old problem with the free space still continues.
Theorically we hace 62 GB free space, but we only can use about 5 GB!

df -kh | grep mail
300G  238G   63G  80% /home/mail

Also, yesterday we have a power outage that force a fsck.ocfs2 of the partition.

We have upgraded to ocfs2 1.4.4 and ocfs2 1.4.3 tools and the kernel version is the same:

rpm -qa | grep ocfs2
ocfs2-2.6.18-128.2.1.el5-1.4.4-1.el5
ocfs2-tools-1.4.3-1.el5
ocfs2console-1.4.3-1.el5

modinfo ocfs2 | grep ^version
version:        1.4.4

uname -r
2.6.18-128.2.1.el5

What we can make? Maybe make a fsck.ocfs2 or mkfs.ocfs2 with the new tools version (1.4.3) ??

Thanks in advance.

Regards,
Pablo







--

------------------------------------------------------------------------

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users

Reply via email to