Yes, but backported and released as ocfs2 1.4 which is yet to be released.
You are on ocfs2 1.2.
Alok Dhir wrote:
I've seen that -- I was under the impression that some of those were
being backported into the release kernels.
Thanks,
Alok
On Feb 7, 2008, at 1:15 PM, Sunil Mushran wrote:
http://oss.oracle.com/projects/ocfs2/dist/documentation/ocfs2-new-features.html
Alok Dhir wrote:
We were indeed using a self-built module due to the lack of an OSS
one for the latest kernel. Thanks for your response, I will test
with the new version.
What are we leaving on the table by not using the latest mainline
kernel?
On Feb 7, 2008, at 12:56 PM, Sunil Mushran wrote:
Are you building ocfs2 with this kernel or are using the ones we
provide for RHEL5?
I am assuming you have built it yourself as we did not release
packages for the latest 2.6.18-53.1.6 kernel till last night.
If you are using your own, then use the one from oss.
If you are using the one from oss, then file a bugzilla with the
full oops trace.
Thanks
Sunil
Alok K. Dhir wrote:
Hello all - we're evaluating OCFS2 in our development environment
to see if it meets our needs.
We're testing it with an iSCSI storage array (Dell MD3000i) and 5
servers running Centos 5.1 (2.6.18-53.1.6.el5xen).
1) Each of the 5 servers is running the Centos 5.1 open-iscsi
initiator, and sees the volumes exposed by the array just fine.
So far so good.
2) Created a volume group using the exposed iscsi volumes and
created a few LVM2 logical volumes.
3) vgscan; vgchange -a y; on all the cluster members. all see the
"md3000vg" volume group. looking good. (we have no intention of
changing the LVM2 configurations much if at all, and can make sure
all such changes are done when the volumes are off-line on all
cluster members, so theoretically this should not be a problem).
4) mkfs.ocfs2 /dev/md3000vg/testvol0 -- works great
5) mount on all Xen dom0 boxes in the cluster, works great.
6) create a VM on one of the cluster members, set up iscsi,
vgscan, md3000vg shows up -- looking good.
7) install ocfs2, 'service o2cb enable', starts up fine. mount
/dev/md3000vg/testvol0, works fine.
** Thanks for making it this far -- this is where is gets interesting
8) run 'iozone' in domU against ocfs2 share - BANG - immediate
kernel panic, repeatable all day long.
"kernel BUG at fs/inode.c"
So my questions:
1) should this work?
2) if not, what should we do differently?
3) currently we're tracking the latest RHEL/Centos 5.1 kernels --
would we have better luck using the latest mainline kernel?
Thanks for any assistance.
Alok Dhir
_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users
_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users
_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users