Always partition the device. Everytime I hear about superblock
corruption, I always notice that that device had not been partitioned.
The mount issue is perplexing. The said mount failed because it found
a _corrupted_ system file. Specifically local_alloc:. Infact it
appears to be all
All,
We are pleased to announce the release of OCFS2 1.2.7-1 for RHEL4
and RHEL5 on i686, x86_64, ppc64 and ia64 architectures.
This release includes three frequently requested updates:
* Fix for bug#5852483. This concerns a possible memory leak in the
size-64 slab during asynchronous io. This
qs is why is the io not completing within 60 secs? You could
try increasing the hb threshold to say 46 (90 secs).
It could be that that while that system may not be doing any
heavy io but some other system using the same storage is.
Has that been ruled out?
In the end, we have to first
Is it a second, as in, just a sec? :)
This could be the result of using the default cluster timeouts
which were fairly low. Refer the Cluster timeout section in the
FAQ for details. You should be able to bump up the disk heartbeat
threshold as recommended. For network timeouts, you will need
to
http://git.kernel.org/?p=linux/kernel/git/torvalds/linux-2.6.git;a=commit;h=7307de80510a70e5e5aa98de1e80ccbb7d90a3a8
Available with 2.6.23.
Christopher Hawkins wrote:
I tried it and had the same issue. Looks like there is no support for shared
writeable mmap, and though a patch was floating
I would recommend 2.6.20+patches at the least.
http://oss.oracle.com/pipermail/ocfs2-users/2007-August/001935.html
[EMAIL PROTECTED] wrote:
Hi,
We're running a default install of Debian 4.
Linux cms1 2.6.18-5-amd64 #1 SMP Tue Oct 2 20:37:02 UTC 2007
x86_64 GNU/Linux
We want to
Could be that those writes are odirect. We avoid updating
the mtime on non-extending odirect writes so as to avoid
a performance hit. This is documented in the ocfs2 faq.
Miguel Molowny Lopez wrote:
Hi all,
we are using OCSF2 in 2 SUSE SLES10 SP1 boxes.
We are running xen on these boxes and we
Actually GFS(2) is also a shared disk clustered filesystem.
Lustre is a good example of a distributed fs.
paul fretter (TOC) wrote:
I had similar confusion myself when first looking for a suitable cluster
FS. I'm not an expert at this, so forgive me if my language appears
simplistic.
There
This is a sysfs issue. Should have nothing to do with ocfs2.
File a bug with RH or Oracle or whomsoever is providing support
for your kernel.
Marcus Alves Grando wrote:
People,
I see one kernel panic in one server of four and i reboot four server
simultaneous. I have only printscreen of ILO2
Randy Ramsdell wrote:
I am not taking sides but I think Alexei's postings are a positive
contribution to this project and more of a contribution than the
lurkers who write nothing to the list. His feedback does have merit
and should be considered valuable although it is critical of ocfs2.
We,
Use network bonding.
Ulf Zimmermann wrote:
-Original Message-
From: [EMAIL PROTECTED] [mailto:ocfs2-users-
[EMAIL PROTECTED] On Behalf Of Alexei_Roudnev
Sent: Thursday, October 11, 2007 11:10
To: Sunil Mushran; Randy Ramsdell
Cc: ocfs2-users@oss.oracle.com
Subject: Re: [Ocfs2-users
so. We switched in Cisco
2960G-48 with Jumbo Frames now and haven't had any reassembly timeouts
since then. Global Cache timeout has gone down significant. Each
Interconnect for Oracle 10G has its own Cisco 2960G-48 now.
-Original Message-
From: Sunil Mushran [mailto:[EMAIL PROTECTED
PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Sunil Mushran
Sent: Tuesday, October 09, 2007 2:27 PM
To: Luis Freitas
Cc: ocfs2-users@oss.oracle.com
Subject: Re: [Ocfs2-users] Cluster setup
Unsure what you mean. If the two servers mount the same
ocfs2 volume and export them via nfs, isn't
BTW, how many bugs have you logged? The least you can do is just that.
Whining will not get you anywhere.
Sunil Mushran wrote:
Alexei_Roudnev wrote:
Does Oracle tests behavior of OCFSv2 in case of:
- 1,000 different users;
YES. By users I assume you mean processes. 500-1000 per node
on a 8
Alexei_Roudnev wrote:
Does Oracle tests behavior of OCFSv2 in case of:
- 1,000 different users;
YES. By users I assume you mean processes. 500-1000 per node
on a 8 node cluster.
- host1 appends to the file and host2 truncate it; then host3 rename file.
YES
- file is removed on node1 but
,
Luis
*/Sunil Mushran [EMAIL PROTECTED]/* wrote:
Appears what you are looking for is a mix of ocfs2 and nfs.
The storage servers mount the shared disks and the reexport
them via nfs to the remaining servers.
ubuntu 6.06 is too old. If you are stuck on Ubuntu LTS, the
next
such as IBRIX.
Brian
-Original Message-
From: [EMAIL PROTECTED]
mailto:[EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Sunil
Mushran
Sent: Tuesday, October 09, 2007 2:27 PM
To: Luis Freitas
Cc: ocfs2-users@oss.oracle.com mailto:ocfs2
Do:
# mounted.ocfs2 -d
Does the volume show up?
Alison Jolley Sanchez wrote:
Everytime I try to start ocfs2 or mount a filesystem, I get the
following error: mount.ocfs2: Bad magic number in superblock while
opening device /dev/emcpowera1
This is a 2 node system, and the other node is
Use network bonding.
Pedro Figueira wrote:
Hello
Is there a way of using more than one network interface for ocfs2 interconnect?
I've searched for this in the documentation but I cannot find any information
about it.
Best regards
Pedro Figueira
Serviço de Estrangeiros e Fronteiras
File a bugzilla at http://oss.oracle.com/bugzilla
Do note the kernel version.
Was this a local mount or a clustered mount?
Add any other info that you feel is relevant.
Any local src changes to the kernel/ocfs2?
Jordi Prats wrote:
HI,
I found this bug. Is this list correct for this type of
Tim Lank wrote:
No problem. :-)
The main points of having a RAC cluster as I understand it are
availability and scalability on low-cost systems. Shouldn't ocfs2 have
the ability to perform online expansion like this? I know that Red Hat's
GFS can add journals to accomodate new nodes while
Please re-read the FAQ entry.
# Is there a limit to the size of an ocfs2 file system?
Yes, current software addresses block numbers with 32 bits. So the file
system device is limited to (2 ^ 32) * blocksize (see mkfs -b). With a
4KB block size this amounts to a 16TB file system. This block
Automatic umount?
The messages do not indicate a umount. How did you detect that the volumes
were umounted?
As in, did you see mount -t ocfs2 or did you also do cat /proc/mounts.
When ocfs2 umounts, it prints the umount message in syslog. I don't
see that message.
Charlie Sharkey wrote:
I
http://oss.oracle.com/bugzilla/show_bug.cgi?id=913
I guess we are missing a __attribute__ ((packed)) in that structure.
I am figuring out why this works on EL4/EL5.
davide rossetti wrote:
On 9/19/07, davide rossetti [EMAIL PROTECTED] wrote:
dear all,
I'm a bit depressed as of now... I
Ahh... this problem is restricted to mainline. We forgot to
pack the vote message while we were pruning it. This problem
is in 2.6.20 onwards.
I'll have the patch fix attached to that bugzilla once I have
tested it.
davide rossetti wrote:
On 9/19/07, davide rossetti [EMAIL PROTECTED] wrote:
http://oss.oracle.com/bugzilla/show_bug.cgi?id=913
Patch fix is attached to the bug. Update the bug with
your results.
Sunil Mushran wrote:
Ahh... this problem is restricted to mainline. We forgot to
pack the vote message while we were pruning it. This problem
is in 2.6.20 onwards.
I'll have
That's the disk heartbeat. 1 block write and X blocks read every 2 secs.
X is the number of nodes defined in the cluster and a block is 512 bytes.
AJ Weber wrote:
I just configured-up my first, two-node OCFS2 cluster with one
mount-point (sdb1). (I know, hooray for you, newbie.)
I'm just
of the same cluster?
I'm running ocfs2-2.6.9-42.0.10.ELlargesmp-1.2.5-2 with
ocfs2-tools-1.2.4-1.
Sunil Mushran wrote:
Probably because you are attempting to run it on a node/cluster
that has the original volume mounted. (The uuid is used as the
dlm domain.)
So rerun the same on a node
merely puts every process accessing the OCFS volume in a D state.
Would you still like me to log another bugzilla issue? I am happy
to do such if you wish.
Cheers,
Matthew
---
Matthew E. Porter
Contegix
Beyond Managed Hosting(r) for Your Enterprise
On Sep 7, 2007, at 12:49 PM, Sunil
Start by diagnosing the the device itself.
W S wrote:
Folks,
We have two linux systems clustered with ocfs2.
Both systems has VERY HIGH % on iowait - ranging from 30% to 60%.
Where should I start looking at first?
Thanks,
--WS
[ocfs2]$ iostat -x
Linux 2.6.9-22.0.1.ELsmp
avg-cpu:
A bugzilla with the oops stack trace will help.
Matthew E. Porter wrote:
Greetings, I am looking for a good way to diagnose random crashes that
are occurring with one of our OCFS clusters. It is a simple 2 node
cluster. debugfs does not seem to indicate any issues.
(Also, I would be happy
What version of the kernel?
Are there any relevant error messages in /var/log/messages.
The error in question is cropping up from o2hb_region_dev_write()
in fs/ocfs2/cluster/heartbeat.c. The messages may give us more
information.
Pierguido wrote:
Hi.
I'm pretty new to ocfs2 and clusters.
I'm
heh... that would be one reason for the error. It had
me puzzled.
Richard Bollinger wrote:
On 8/24/07, Sunil Mushran [EMAIL PROTECTED] wrote:
File a bugzilla. It will be easier to track the issue.
After a fresh mkfs.ocfs2, do:
# debugfs.ocfs2 -R dump //slot_map /tmp/slotmap /dev/sdX
Read
Please file a bugzilla. It is very hard to track issue via email.
Attach the trace below. You should also see a corresponding
message in one of the other nodes. Specifically node 0. Add
that too in the bugzilla.
Daniel wrote:
Hello
I'm still having weekly panics on my system, but now I've at
That's not good. Does the device have enough free space?
File a bugzilla with the details already provided. Also attach the
output of the stat_sysdir.sh script.
http://oss.oracle.com/~smushran/.debug/scripts/stat_sysdir.sh
Christophe Bouder wrote:
hello,
on my system debian 64,vanilla kernel
outputs)
top
(1 screen)
slabtop (or equivalent, I dont remember)
- Original Message -
*From:* Arnold Maderthaner
mailto:[EMAIL PROTECTED]
*To:* Sunil Mushran mailto:[EMAIL PROTECTED
cluster.conf should be exactly the same on all three nodes.
While private ip addresses should be used, public will also
work as long as the addresses are the same in all cluster.confs.
Also, o2cb_ctl is a bit picky as to the format of the file. As in,
it expects a newline between stanzas and the
You could be encountering Novell bugzilla 296606. It is specific
to SLES10 (and SP1). Novell owns the bug.
Sebastian Reitenbach wrote:
Hi,
I am on SLES 10, SP1, x86_64, running the distribution rpm's of ocfs:
ocfs2console-1.2.3-0.7
ocfs2-tools-1.2.3-0.7
I have a two node ocfs2 cluster
File a bugzilla. It will be easier to track the issue.
After a fresh mkfs.ocfs2, do:
# debugfs.ocfs2 -R dump //slot_map /tmp/slotmap /dev/sdX
Read /tmp/slotmap in a binary editor. The first 1K should be all FF's.
Attach the file to the bugzilla.
Richard Bollinger wrote:
On 8/22/07, Sunil
That's fine. The only limitation is how the file is updated.
In 1.2, one cannot create a shared writeable mmap. But one
is free to use write() and have shared readonly mmaps.
Patch to address shared writeable mmap has been checked into
mainline. It will be available in 2.6.23 and with ocfs2 1.4.
All,
We are pleased to announce the release of OCFS2 1.2.5-6 for EL4 and
OCFS2 1.2.6-6 for EL5. Refer news for the full list of changes in the
the two releases.
http://oss.oracle.com/projects/ocfs2/news/article_14.html
http://oss.oracle.com/projects/ocfs2/news/article_15.html
Please note that
Preferably file a bugzilla with the traces (var/log/messages) and also
the information you have already posted.
On Tue, Aug 21, 2007 at 06:49:30AM -0400, Richard Bollinger wrote:
On 8/21/07, Sunil Mushran [EMAIL PROTECTED] wrote:
Do:
# debugfs.ocfs2 -R stats /dev/sdX
Ensure the Number
Could you describe the restart problem in more detail.
As far as the second issue goes, is there some multipathing in play.
I am unclear how a fs can be responsible for that issue.
On Wed, Aug 22, 2007 at 12:29:13PM +0200, ludek_coufal wrote:
Hello,
configuration:
- 4 HP Proliant 460c servers
Repeat the first test. This time run top on the first server. Which
process is eating the cpu? From the time it appears that the second
server is waiting for the AST that the first server is slow to send. And
it could be slow because it may be busy flushing the data to disk.
How much memory/cpu do
by Oracle or someone else?
On Wed, Aug 22, 2007 at 08:49:30AM -0400, Richard Bollinger wrote:
On 8/22/07, Sunil Mushran [EMAIL PROTECTED] wrote:
Do:
# debugfs.ocfs2 -R slotmap /dev/sdX
slotmap maintains the slot-to-nodenum mappings.
# debugfs.ocfs2 -R slotmap /dev/sde1
Slot# Node
Can you use the packages provided on oss.oracle.com. The issue is
strange and I would like to rule out a build error.
On Wed, Aug 22, 2007 at 02:49:38PM -0400, Richard Bollinger wrote:
On 8/22/07, Sunil Mushran [EMAIL PROTECTED] wrote:
Which mkfs.ocfs2 are you using? Actually the toolset looks
Do:
# debugfs.ocfs2 -R stats /dev/sdX
Ensure the Number of Slots is what it is supposed to be.
Also, do:
# debugfs.ocfs2 -R ls -l // /dev/sdX
Ensure you see the journals for all the above slots.
On Mon, Aug 20, 2007 at 03:10:50PM -0400, Richard Bollinger wrote:
Had to recreate one of eight
Please file a bugzilla on oss.oracle.com/bugzilla. It's easier
to keep track of issues that-a-way.
Attach the messages file from all nodes in the cluster. While
the logs you have provided should be enough, having the complete
logs is better as it provides a fuller picture.
Daniel wrote:
Hello
on how long a failure like this
could last?
Thanks for your help.
Andrew
On Mon, 2007-07-23 at 13:37 -0700, Sunil Mushran wrote:
Yes, the failure is temporary.
Andrew D. Ball wrote:
On Tue, 2007-07-17 at 16:17 -0700, Sunil Mushran wrote:
Ahh... this is a 1.2 feature
through alternatives.
Peace,
Andrew
On Sat, 2007-07-21 at 15:44 -0700, Sunil Mushran wrote:
No. Adding a node will not help.It's a node death during the said
operation that is causing the temporary failure.
I am curious as to why is this such big issue for you. Can you elaborate
It should behavw as you expect it to. That's the idea.
What are the errors when mkdir fails?
As in, userspace and dmesg.
Andrew D. Ball wrote:
I would really like to see the following behavior:
(1) I start with a two-node cluster, both nodes online, with an ocfs2
filesystem mounted on both
, 2007-07-17 at 15:31 -0700, Sunil Mushran wrote:
It should behavw as you expect it to. That's the idea.
What are the errors when mkdir fails?
As in, userspace and dmesg.
Andrew D. Ball wrote:
I would really like to see the following behavior:
(1) I start with a two-node cluster, both nodes
Use the build command as listed on the ocfs2 home page.
On Fri, Jul 13, 2007 at 10:54:20AM -0400, Randy Ramsdell wrote:
Any ideas on this crash? The system remained available for login, but
basically hosed all the other nodes in the cluster. It is interesting
that the other nodes did not write
This should be totally transparent the the filesystem.
Use ls -i to debug. That will dump the inode number.
As the inode number is tied to the block number on the device,
see if they match.
You can also use debugfs.ocfs2 to traverse thru the directories
on the block devices directly.
Andrew D.
FAQ has information on this.
Ulf Zimmermann wrote:
I looked around, found older post which seems not applicable anymore. I
have a cluster of 2 nodes right now, which has 3 OCFS2 file systems. All
the file systems were formatted with 4 node slots. I added the two news
nodes (by hand, by
File a bugzilla for the first issue. The second issue is not related to
ocfs2.
Append the messages file of this node and node 0. Add whatever else you
deem relevant to the issue activity on the server/cluster, etc.
Daniel wrote:
Hello
System: Two brand new Dell 1950 servers with dual
If it is reproducible at will, please can you rerun that test with
tracing enabled.
# debugfs.ocfs2 -l ENTRY EXIT INODE allow
do test...
# debugfs.ocfs2 -l ENTRY EXIT deny INODE off
File a bug in oss.oracle.com/bugzilla and attach the log trace. Entire
message file will do.
Do mention
All,
We are pleased to announce the release of OCFS2 1.2.5-2 for EL4.
This release has one bug fix as listed in the news.
http://oss.oracle.com/projects/ocfs2/news/article_13.html
The nature of this bug is such that it mainly affects Oracle RDBMS
users. Almost all users who have encountered
-0700 schrieb Sunil Mushran:
It's not the bandwidth but the latency that is the issue.
Try the same with gige.
Philipp Wehrheim wrote:
Am Donnerstag, den 14.06.2007, 16:08 -0700 schrieb Sunil Mushran:
Did you try using a gige interconnect?
No not yet.
Do you think
service while starting heartbeat
mount.ocfs2: Error when attempting to run /sbin/ocfs2_hb_ctl: Operation
not permitted
There isn't any good installation guide? All I've found googling are
incomplete
Thanks!
Jordi
Sunil Mushran wrote:
Are you applying that patch to a plain 2.6.20?
Or, did you
Start with the hardware, specifically the shared disk. See
if the performance of concurrent r/ws from two nodes is slower
than from any one node. You can run hdparm -t concurrently
from the two nodes. Or if you have no data on the disks, use a
destructive variation of the same.
The other thing
Do you have netdump/netconsole setup? The actual reason
for the reboot will only be caught in those logs.
[EMAIL PROTECTED] wrote:
On Thursday night I had one of my Linux ( RHEL v4 2.6.9-55.ELsmp
ocfs2-2.6.9-55.ELsmp-1.2.5-1) reboot 2X. Once at 21:46 then again at 22:26.
The only clues I have
All,
We are pleased to announce the release of OCFS2 1.2.6 and OCFS2 Tools
1.2.6 for EL5.
Apart from the changes required to build OCFS2 on EL5, the main difference
between this and the previous release relates to the default cluster
timeouts.
We have updated the defaults in this release.
The default disk timeout in ocfs2 is fairly low. It just could
be that that needs to be bumped up.
Do you have the netconsole output of the crash?
Jeff Bachtel wrote:
We are sharing ocfs2 volumes to Xen domU processes on our SAN. From a
stability standpoint on SLES10 (not SP1 yet) and OpenSuSE
Did you try using a gige interconnect?
Philipp Wehrheim wrote:
Hi everbody,
I've just created a test setup for my company with two pc's drbd (0.8.3)
primary/primary mode with OCFS2 on top of it. the pc's are running suse
10.2 with kernel 2.6.21.2. the Hardware CPU ~1GHz ~12 MBRam and one NIC
Do:
# fsck.ocfs2 -f /dev/sda1
Rich Amick wrote:
Testing OCFS2 on a 2 node cluster.
Both nodes are HP DL380G5 with RHEL4U3 x86_64 (2.6.9-34.ELsmp).
OCFS2 version 1.2.5-1.
Shared storage is MSA1000.
Formatted a 142GB volume with:
mkfs.ocfs2 -b 4K -C 32K -N 4 -L u01 /dev/sda1
Mounted the
Check whether it is the same as in 2.6.21. We run continuous tests against
mainline kernels and we've not encountered these problems.
The patch fix atop 2.6.21 is as follows:
http://www.kernel.org/pub/linux/kernel/people/mfasheh/ocfs2/backports/2.6.21/
The break up of the above can be viewed
Network timeout.
Did a node die during the delete? If so, then the error is
to be expected. As in, retry the rm and it should succeed.
Or, were you umounting that volume on another node during
the said delete? If so, then this should not happen.
[EMAIL PROTECTED] wrote:
Hi,
I am on SLES
This smells like a uuid clash. But it could also be a bug. :)
Do:
# mount -t debugfs debugfs /debug (if not already mounted)
# ls -R /debug
In short, the mount is complaining that a entry with the
same uuid already exists in that location.
If that is the case, do:
# cat /proc/fs/ocfs2/*/uuid
Woods
Sunil Mushran
[EMAIL PROTECTED]
acle.com
Or we could provide 1.2.4 for 2.6.9-55. If you still have this
issue, we could always do that. Let me know.
[EMAIL PROTECTED] wrote:
I am in the middle of rolling RHEL 4 update 5 ( 2.6.9-55) through my
servers. We use a development, QA, and production rollout.
Unfortunately I have ocfs members
Yassine
On 5/29/07, *Sunil Mushran* [EMAIL PROTECTED]
mailto:[EMAIL PROTECTED] wrote:
The debugfs module is bundled with the ocfs2 packages for 2.6.9
(rhel4).
Appears as if the module is not getting loaded.
Do:
# rpm -qf /lib/modules/`uname -r`/kernel/fs/debugfs/debugfs.ko
the firewire module I did install the centosplus one.
That may explain why.
[EMAIL PROTECTED] opt]$ rpm -qa | grep ocfs2
ocfs2console-1.2.4-1
ocfs2-tools-1.2.4-1
ocfs2-2.6.9-55.ELsmp-1.2.5-1
Yassine
On 5/29/07, * Sunil Mushran* [EMAIL PROTECTED]
mailto:[EMAIL PROTECTED] wrote:
debugfs
disc 3
Kendall, Kim wrote:
Which CD is it on, it didn't install by itself. (We're trying it before
we purchase / register it)
-Original Message-
From: Sunil Mushran [mailto:[EMAIL PROTECTED]
Sent: Wednesday, May 23, 2007 12:26 PM
To: Kendall, Kim
Cc: ocfs2-users@oss.oracle.com
Such issues are handled best via bugzilla. File one
on oss.oracle.com/bugzilla with all the details.
The most important detail would be node3's netdump
or netconsole output. The real reason for the outage
will be in that dump.
Marcus Alves Grando wrote:
Hi list,
Today i have a problem with
spfile needs to be on a shared storage for RAC.
So it can be stored on an ocfs2 volume.
The 128K clustersize is a recommendation.
mount options datavolume and nointr are a requirement.
Wu, Hai wrote:
I think I saw somewhere it says, to put spfile in ocfs2 partition, it should
use 128k cluster
reproduce that? First i need more
info how reproduce that.
Regards
Sunil Mushran wrote:
Such issues are handled best via bugzilla. File one
on oss.oracle.com/bugzilla with all the details.
The most important detail would be node3's netdump
or netconsole output. The real reason for the outage
OCFS2 for EL4 is shipped alongwith the iso. If registered,
you can use up2date, as listed on oss.oracle.com/projects/ocfs2
to get the package too.
Kendall, Kim wrote:
Installed Oracle Enterprise Linux R4 Update 4
Kernel is: 2.6.9-42.0.0.0.1.ELsmp
What version of ocfs2 will load in it? I tried
That should work.
Kendall, Kim wrote:
What we will be doing it creating EMC Clones of the older version of
OCFS2 volumes, then mounting the clone volumes to another host that has
a newer version of OCFS2.
-Original Message-
From: Mark Fasheh [mailto:[EMAIL PROTECTED]
Sent: Monday,
Ensure you have mounted the volume storing the
ocr and the voting diskfile with the nointr,datavolume
mount options.
http://oss.oracle.com/projects/ocfs2/dist/documentation/ocfs2_faq.html#RAC
Dheerendar Srivastav wrote:
Dear sir ,
We have used RHEL 4.0 with kernel 2.6.9-42.0.2.ELsmp with
will be
killed.
Without intr, if your NFS server died, you wil not be able to shutdown your
server without power reset (or reboot -n -f in some cases).
- Original Message -
From: Sunil Mushran [EMAIL PROTECTED]
To: Dheerendar Srivastav [EMAIL PROTECTED]
Cc: Ocfs2-users@oss.oracle.com
Sent: Monday
then power reset the whole box.
Do you want application (oracle archiver for example) to be sticked forever
if one of NFS servers (one of a few archive destinations) don't answer? I
dont think so,
- Original Message -
From: Sunil Mushran [EMAIL PROTECTED]
To: Alexei_Roudnev [EMAIL PROTECTED]
Cc
Do you have a netdump/netconsole server configured?
If so, get the oops stack trace. That will tell us the reason
for the reboot.
Gaetano Giunta wrote:
Hello.
On a 2 node RAC 10.2.0.3 setup, on RH ES 4.4 x86_64, with ocfs 1.2.5-1, we are
experiencing some troubles with RMAN: when the archive
That the datavolume mount option is not on ocfs2 shipped alongwith
the mainline kernel should not come as a surprise to anyone. The
mount option is only present in distributions that the Oracle RDBMS
supports. The option is very specific to Oracle and we are exploring
getting rid of it.
Alexei,
This is a very old release. Use a more recent 1.2 release. 1.2.5 is the
latest and we have modules for 2.6.9-22.
Secondly, as you device is only 1G, you may want to tune down the
clustersize to 4K.
# mkfs.ocfs2 -b 4K -C 4K -L label /dev/sdb1
If that does not work, strace it.
# strace -ff -o
o2cb is the cluster stack. The ocfs2 volume also needs to be mounted.
Is the volume mounted on all nodes?
Dheerendar Srivastav wrote:
Hi
We are installation of oracle RAC and install the ocfs2 but check
/o2cb status is running fine but when we installation of the crs the
error message show
None of what you have written allows you to use our resources to
spread your opinions as official recommendation.
Alexei_Roudnev wrote:
Yes, you distribute and support OCFSv2; but Oracle is another story, and the
common recommendation
from many different Oracle engineers (not related to OCFSv2)
Ulf Zimmermann wrote:
Just to add some comments from a user of Oracle 9i with OCFSv1 on RedHat
AS2.1 who tried to upgrade to EL4 and OCFSv2 and failed miserable:
Oracle support pretty much told us the problems we were running into are
problems of OCFSv2 and they weren't really willing to help
Please do. Provide the kernel/ocfs2 version. Also, I
had mentioned other things in my previous email like
scanlocks, etc. Also, attach complete outputs. Easier
to read raw data. And attach all the outputs.
Stephan Hendl wrote:
After 9 days without an error we got a further webserver hang last
Userspace hb is a suse extension to ocfs2. Only available on sles10.
Chakravarti, Avishek (STSD) wrote:
Hi,
Is it possible to configure ocfs2 1.2.5 to use user-space heartbeat on
Redhat Enterprise Linux?
I was previously using ocfs2 1.2.3 on SLES10 and had configured
Heartbeat2.0 to do
strace both processes. The answer will be in there.
On Mon, 2007-04-30 at 17:59 +0200, Gaetano Giunta wrote:
we recently deployed ocfs2 (rel 1.2.3) on a RH ES 4 update 4, x86_64
the ocfs2 partition is being mounted at the same time by two nodes
while using cp to move data between local
This is a build problem.
From ocfs2 home page:
Other Distributions: If your distribution does not include OCFS2,
you will need to download and build the sources. To build against
the mainline kernel, download the kernel from
It should have encountered this problem too.
On Mon, 2007-04-30 at 15:53 -0400, Randy Ramsdell wrote:
Sunil Mushran wrote:
This is a build problem.
From ocfs2 home page:
Other Distributions: If your distribution does
We have LP1 in our test machines.
Is the instability in your system related to fencing?
If so, update to OCFS2 1.2.5 and use more relaxed timeouts as
listed in the FAQ.
On Fri, 2007-04-27 at 14:50 -0400, Mike McCoy wrote:
Does anyone have any experience with Emulex HBA cards (LP1)
# debugfs.ocfs2 -R stats -h /dev/sdy2 | grep Cluster Size
Block Size Bits: 12 Cluster Size Bits: 17
12 = 4K
17 = 128K
Have you tried stracing the process?
# strace -tt -T -o /tmp/strace.out ...
Ulf Zimmermann wrote:
Is there a way to see how a file system was formatted, i.e. the
I am unclear as to the exact issue. As in, are process(es) on a
particular node hanging when accessing (a) ocfs2 volume or is the
problem cluster-wide. As in, whether processes on all nodes are
hanging.
Let me list some steps that will help you gather the relevant info.
Start with downloading
What's the blocksize?
Ulf Zimmermann wrote:
-Original Message-
From: Sunil Mushran [mailto:[EMAIL PROTECTED]
Sent: 04/25/2007 10:31
To: Ulf Zimmermann
Cc: ocfs2-users@oss.oracle.com
Subject: Re: [Ocfs2-users] Some questions about ocfs2
# debugfs.ocfs2 -R stats -h /dev/sdy2 | grep
:
-Original Message-
From: Sunil Mushran [mailto:[EMAIL PROTECTED]
Sent: 04/25/2007 12:16
To: Ulf Zimmermann
Cc: ocfs2-users@oss.oracle.com
Subject: Re: [Ocfs2-users] Some questions about ocfs2
What's the blocksize?
Block Size Bits: 12 Cluster Size Bits: 14
Ulf Zimmermann wrote
#define EADDRNOTAVAIL 99 /* Cannot assign requested address */
You will need to make the appropriate changes to the
ip address in /etc/ocfs2/cluster.conf too. Ensure the config
file is the same across the cluster.
Eduardo Paixao Niel wrote:
Hi,
We have an environment with 2 nodes
):ocfs2_dlm_init:2007 ERROR: status = -107
(5886,0):ocfs2_mount_volume:1062 ERROR: status = -107
ocfs2: Unmounting device (8,17) on (node 1)
Thanks,
Eduardo
-Mensagem original-
De: Sunil Mushran [mailto:[EMAIL PROTECTED]
Enviada em: terça-feira, 24 de abril de 2007 16:29
Para: Eduardo
701 - 800 of 943 matches
Mail list logo