That might because you have configured user mode dlm. What does |cat
/sys/o2cb/heartbeat_mode show, user?. If so run, /etc/init.d/o2cb
configure and answer n for the following
||Use user-space driven heartbeat? (y/n) [y] n
thanks,
--Srini
||
|Andy Johnson wrote:
Hello,
I have
things:
a) Recover orphans during mount of the slot that it is using.
b) Recover orphans in all offline slots during recovery.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/journal.c | 44 +++-
1 files changed, 27 insertions(+), 17
are
clean they will not queue to recover their orphan directory.
This patch queues to recover orphans when the slot is next used.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/journal.c | 21 -
1 files changed, 8 insertions(+), 13 deletions(-)
diff --git
Next 3 patches does the following
1) movies ocfs2_slot_info struct from slot_map.c to slot_map.h
2) patch to recover orphans during mount even if the journal is clean
3) patch to recovery orphans in offline slots
___
Ocfs2-devel mailing list
recovers it's own slot, which
leaves orphans in offline slots.
This patch queues complete_recovery to clean orphans for all offline slots
during mount and node recovery.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/journal.c | 106
recovers it's own slot, which
leaves orphans in offline slots.
This patch queues complete_recovery to clean orphans for all offline slots
during mount and node recovery.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/journal.c | 136
recovers it's own slot, which
leaves orphans in offline slots.
This patch queues complete_recovery to clean orphans for all offline slots
during mount and node recovery.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
Signed-off-by: Joel Becker joel.bec...@oracle.com
---
fs/ocfs2/journal.c | 140
The following patch is a backport of patch that recovers orphans from offline
slots. It is being backported from mainline to 1.4
mainline patch: 0001-Patch-to-recover-orphans-in-offline-slots-during-rec.patch
Thanks,
--Srini
___
Ocfs2-devel mailing
hmm, even if we queue the orphan recovery, inode may not get cleaned if
the inode is still around on some node right? The node where the inode
is still cached will vote no again?
Sunil Mushran wrote:
Joel Becker wrote:
Srini,
Ok, you can go ahead and cook up the background orphan
did you use -o datavolume, nointr options for mounting?
keyur patel wrote:
Hello All,
I have installed Oracle Cluster Manager on linux x86-64 nit. I am
using ocfs file system for quorum file. But I am getting following
error. Please see ocfs configureation below. I would appreciate, if
-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlmglue.c | 58 +
fs/ocfs2/dlmglue.h |8 +++
fs/ocfs2/journal.c | 109 +++
fs/ocfs2/journal.h | 12 +
fs/ocfs2/ocfs2.h|2 +
fs
at a time. It is done once every X seconds, where X is a value between
ORPHAN_SCAN_SCHEDULE_TIMEOUT/2 and ORPHAN_SCAN_SCHEDULE_TIMEOUT milliseconds.
Each time the scan is done by different node so eventually the node that has the
inode cached will get to wipe the file.
Signed-off-by: Srinivas Eeda
Patch to track delayed orphan scan timer statistics.
Modifies ocfs2_osb_dump to print the following:
Orphan Scan= Local: 10 Global: 21 Last Scan: 67 seconds ago
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/journal.c |2 ++
fs/ocfs2/ocfs2.h |4 +++-
fs/ocfs2
at a time. It is done once every X seconds, where X is a value between
ORPHAN_SCAN_SCHEDULE_TIMEOUT/2 and ORPHAN_SCAN_SCHEDULE_TIMEOUT milliseconds.
Each time the scan is done by different node so eventually the node that has the
inode cached will get to wipe the file.
Signed-off-by: Srinivas Eeda
Resending after adding another patch to display delayed orphan scan statistics.
___
Ocfs2-devel mailing list
Ocfs2-devel@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-devel
the cluster so that no one node has to perform the task all the
time.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlmglue.c | 47 +
fs/ocfs2/dlmglue.h | 11 +
fs/ocfs2/journal.c | 106 +++
fs
Patch to track delayed orphan scan timer statistics.
Modifies ocfs2_osb_dump to print the following:
Orphan Scan= Local: 10 Global: 21 Last Scan: 67 seconds ago
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/journal.c |2 ++
fs/ocfs2/ocfs2.h |4 +++-
fs/ocfs2
Resending after implementing review comments.
___
Ocfs2-devel mailing list
Ocfs2-devel@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-devel
Patch to track delayed orphan scan timer statistics.
Modifies ocfs2_osb_dump to print the following:
Orphan Scan= Local: 10 Global: 21 Last Scan: 67 seconds ago
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
Signed-off-by: Sunil Mushran sunil.mush...@oracle.com
---
fs/ocfs2/journal.c
Resending after implementing review comments.
___
Ocfs2-devel mailing list
Ocfs2-devel@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-devel
such inodes. Care has been taken to distribute the
workload across the cluster so that no one node has to perform the task all the
time.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlmglue.c | 51 ++
fs/ocfs2/dlmglue.h | 10
fs/ocfs2
Next two patches are backport of orphan scan timer patches to ocfs2-1.4
___
Ocfs2-devel mailing list
Ocfs2-devel@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-devel
in ocfs2_super when reflink is
ongoing(I will do it).
Make sense?
Yes, I can restrict the node to recover it's own and offline slots. I
can make the node to recover it's own slot every time the timer fires
and offline slots in round robin way(current way)
Regards,
Tao
Srinivas Eeda wrote
Tao Ma wrote:
Hi Joel,
This reply may be really too late. :)
Joel Becker wrote:
On Wed, Jun 10, 2009 at 01:37:53PM +0800, Tao Ma wrote:
I also have some thoughts for it. Wish it isn't too late.
Well, if we come up with changes it will affect what I push, but
that's OK.
is cancelled and messaging resumes.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/cluster/tcp.c | 94 +++
fs/ocfs2/cluster/tcp_internal.h |9 ++--
2 files changed, 60 insertions(+), 43 deletions(-)
diff --git a/fs/ocfs2
When a file system is mounted local, it may be enough to remount it read only
on seeing corruptions.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/super.c | 10 ++
1 files changed, 6 insertions(+), 4 deletions(-)
diff --git a/fs/ocfs2/super.c b/fs/ocfs2/super.c
messages
to/from evicted node. If network connection comes back before the eviction,
quorum decision is cancelled and messaging resumes.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/cluster/tcp.c | 69 +++
fs/ocfs2/cluster
This patch rollbacks earlier fix that tries to re-establish network connection
when network timeout happens. Reconnect was re-cycling sockets which results
in lost messages resulting in hangs.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/cluster/tcp.c | 50
Following 3 patches fixes:
1. rollback's reconnect fix
2. delay enotconn for sends, receives till a node reconnects/dies after a
lost connection.
3. Correct's keepalive protocol
Thanks,
--Srini
___
Ocfs2-devel mailing list
they received. So
nodes with this patch will always receive a response message.
So, in a mixed setup, both nodes will always hear the heartbeat from
each other :).
thanks,
--Srini
Joel Becker wrote:
On Thu, Jan 28, 2010 at 08:51:11PM -0800, Srinivas Eeda wrote:
case
No harm, just doubles heartbeat messages which is not required at all.
Sunil Mushran wrote:
What's the harm in leaving it in?
Srinivas Eeda wrote:
Each node that has this patch would send a O2NET_MSG_KEEP_REQ_MAGIC
every 2 seconds(default). So, nodes without this patch would always
alive every 2 seconds.
Sunil Mushran wrote:
How will it double? The node will send a keepalive only if it has
not heard from the other node for 2 secs.
Srinivas Eeda wrote:
No harm, just doubles heartbeat messages which is not required at all.
Sunil Mushran wrote:
What's the harm in leaving
.
As in, not wait for the response to requeue. But we'll still be smart
about
it in the sense that not send a hb even if the nodes are communicating
otherwise.
Srinivas Eeda wrote:
In old code a node cancels and re queues keep alive message when it
hears from the other node. If it didn't hear in 2
Yea, they don't expect/wait for a response for keep alive message.
On 2/17/2010 5:49 PM, Joel Becker wrote:
On Wed, Feb 17, 2010 at 10:24:30AM -0800, Srinivas Eeda wrote:
Each node that has this patch would send a O2NET_MSG_KEEP_REQ_MAGIC
every 2 seconds(default). So, nodes without
delivery. However the intention of this feature was to send
a keepalive message every timeout seconds. This patch sends a message for
every keepalive time interval.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/cluster/tcp.c |6 +-
1 files changed, 5 insertions(+), 1
DLM_ASSERT_RESPONSE_MASTERY_REF) which creates a
hole that results in loss of refmap bit on the master node.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmmaster.c |4 +---
1 files changed, 1 insertions(+), 3 deletions(-)
diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm
Sunil, Joel, thanks for modifying the comments :)
On 3/22/2010 6:47 PM, Joel Becker wrote:
On Mon, Mar 22, 2010 at 06:20:32PM -0700, Sunil Mushran wrote:
yes, your wording is better. and yes, dlm-spinlock is the
top level lock.
This patch is now in the 'fixes' branch of
informing the
master directly. This is easily fixed by holding the dlm spinlock a
little longer in the mastery handler.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmmaster.c |4 +---
1 files changed, 1 insertions(+), 3 deletions(-)
diff --git a/fs/ocfs2/dlm
This patch logs socket state changes that lead to socket shutdown.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/cluster/tcp.c |3 +++
1 files changed, 3 insertions(+), 0 deletions(-)
diff --git a/fs/ocfs2/cluster/tcp.c b/fs/ocfs2/cluster/tcp.c
index 334f231..6d0d228
This patch logs socket state changes that lead to socket shutdown.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/cluster/tcp.c |3 +++
1 files changed, 3 insertions(+), 0 deletions(-)
diff --git a/fs/ocfs2/cluster/tcp.c b/fs/ocfs2/cluster/tcp.c
index 334f231..8bda1ea
The following patch logs socket shutdown messages. Below is the snippet of how
the message looks (new message ends with ... shutdown, state #)
[r...@el532p-3 ~]# mount /dev/hdb /vol1
Mar 31 11:14:18 el532p-3 kernel: connection to node el532p-2 (num 64) at
10.35.70.104: shutdown, state 8
Mar
thanks for doing this patch. I have a little comment, wondering if there
could be a window where node B sent the lock info to node C as part of
recovery and removed flag DLM_LOCK_RES_RECOVERING while dlm_thread was
still purging it. In that case dlm_thread will still continue to remove
it from
On 5/24/2010 7:50 PM, Wengang Wang wrote:
delay deref message if DLM_LOCK_RES_RECOVERING is set (which means
recovery got to the lockres before dlm_thread could), move the
lockres to the end of the purgelist and retry later.
If you meant checking before sending DEREF, it could cause a
Comments inline
On 6/3/2010 9:37 AM, Wengang Wang wrote:
Changes to V1:
1 move the msleep to the second runs when the lockres is in recovery so the
purging work on other lockres' can go.
2 do not inform recovery master if DLM_LOCK_RES_DROPPING_REF is set and don't
resend deref in this
On 6/3/2010 6:43 PM, Wengang Wang wrote:
Srini,
On 10-06-03 18:06, Srinivas Eeda wrote:
Comments inline
On 6/3/2010 9:37 AM, Wengang Wang wrote:
Changes to V1:
1 move the msleep to the second runs when the lockres is in recovery so the
purging work on other lockres' can go.
2 do
On 6/3/2010 10:37 PM, Wengang Wang wrote:
Srini,
On 10-06-03 19:17, Srinivas Eeda wrote:
Can you please explain the idea of the new flag
DLM_LOCK_RES_DE_DROP_REF :)
If the idea of the fix is to address the race between purging and
recovery, I am wondering DLM_LOCK_RES_DROPPING_REF
.
The fix is to add new state DLM_LOCK_RES_IN_USE which would prevent
dlm_thread from purging the lockres and/or unhashing lockres.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmcommon.h |1 +
fs/ocfs2/dlm/dlmlock.c |4
fs/ocfs2/dlm/dlmmaster.c |5
Wengang, thanks for the patch. My comments are inline :)
On 6/8/2010 7:38 AM, Wengang Wang wrote:
dlm_thread(when purges a lockres) races with another thread that is running on
dlmlock_remote(). dlmlock_remote() can add a lock to the blocked list of the
lockres without taking dlm-spinlock.
Please ignore this patch. I'll resend this patch along with dlm purge
lockres for completeness.
On 6/9/2010 7:07 PM, Srinivas Eeda wrote:
dlm_thread sends a deref message to the master node. At the same time,
another thread sends a new lock request to the master node.
Since dlm_thread
from the master lockres is not found
This patch marks the lockres with a new state DLM_LOCK_RES_IN_USE which would
protect lockres from dlm_thread purging it.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmcommon.h |1 +
fs/ocfs2/dlm/dlmlock.c |4
fs/ocfs2
The lock order in this code causes dead lock, not caused by your patch.
The lock order in dlm_query_join_handler is
dlm_domain_lock
-dlm-spinlock
dead locks with ..
dlm_lockres_put calls dlm_lockres_release while holding dlm-spinlock
which calls dlm_put which gets dlm_domain_lock. So the
patch looks good, it fixes the umount code path which prevents a lockres
from migrating if it needs to be recovered. I have few comments on the
scenario you described.
On 6/11/2010 3:25 AM, Wengang Wang wrote:
Any comment on this patch?
regards,
wengang.
On 10-05-25 15:59, Wengang Wang
On 6/17/2010 1:32 AM, Joel Becker wrote:
On Wed, Jun 16, 2010 at 06:44:43PM -0700, Sunil Mushran wrote:
One way to skip a lockres in the purgelist is to list_del_init() and
list_add_tail(). That simplifies the patch a lot.
I have attached a quick dirty patch. See if that satisfies all the
On 6/17/2010 7:48 AM, Sunil Mushran wrote:
On 06/17/2010 01:35 AM, Srinivas Eeda wrote:
On 6/17/2010 1:32 AM, Joel Becker wrote:
On Wed, Jun 16, 2010 at 06:44:43PM -0700, Sunil Mushran wrote:
One way to skip a lockres in the purgelist is to list_del_init() and
list_add_tail
Sunil,
as of now, there is still a window in dlm_get_lock_resource, where it
finds the lockres but it doesn't protect it from getting purged. Second
patch fixes this by marking it in_use, can you please review that one as
well.
Thanks,
--Srini
On 6/17/2010 8:06 AM, Sunil Mushran wrote:
On
is adding
more messiness to an area that is already very messy.
Sunil
On 06/15/2010 09:43 PM, Srinivas Eeda wrote:
This patch fixes the following hole.
dlmlock tries to create a new lock on a lockres that is on purge
list. It calls
dlm_get_lockresource and later adds a lock to blocked list
it from getting reused.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmthread.c | 55 +
1 files changed, 26 insertions(+), 29 deletions(-)
diff --git a/fs/ocfs2/dlm/dlmthread.c b/fs/ocfs2/dlm/dlmthread.c
index 11a6d1f..79d1ef6
it from getting reused.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmthread.c | 76 --
1 files changed, 33 insertions(+), 43 deletions(-)
diff --git a/fs/ocfs2/dlm/dlmthread.c b/fs/ocfs2/dlm/dlmthread.c
index 11a6d1f..cb74689
it from getting reused.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmthread.c | 79 +++--
1 files changed, 33 insertions(+), 46 deletions(-)
diff --git a/fs/ocfs2/dlm/dlmthread.c b/fs/ocfs2/dlm/dlmthread.c
index 11a6d1f..6822f9a
Sunil, Joel, Wengang. Thanks for reviewing the patch and your comments.
On 6/23/2010 10:00 AM, Sunil Mushran wrote:
Signed-off-by: Sunil Mushransunil.mush...@oracle.com
On 06/22/2010 10:48 PM, Srinivas Eeda wrote:
There are two problems in dlm_run_purgelist
1. If a lockres is found
Joel Becker wrote:
On Tue, Jun 22, 2010 at 10:48:28PM -0700, Srinivas Eeda wrote:
+if (!__dlm_lockres_unused) {
+mlog(ML_ERROR, found lockres %s:%.*s: in use after deref\n,
+ dlm-name, res-lockname.len, res-lockname.name
it from getting reused.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
Signed-off-by: Sunil Mushransunil.mush...@oracle.com
---
fs/ocfs2/dlm/dlmthread.c | 79 +++--
1 files changed, 33 insertions(+), 46 deletions(-)
diff --git a/fs/ocfs2/dlm
protecting it from getting reused.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
Acked-by: Sunil Mushran sunil.mush...@oracle.com
---
fs/ocfs2/dlm/dlmthread.c | 80 +++--
1 files changed, 34 insertions(+), 46 deletions(-)
diff --git a/fs/ocfs2/dlm
thanks for making this patch, it looks good just few minor changes about
comments
On 7/23/2010 5:15 AM, Wengang Wang wrote:
In the following situation, there remains an incorrect bit in refmap on the
recovery master. Finally the recovery master will fail at purging the lockres
due to the
on master list after the node sent
DLM_EXIT_DOMAIN_MSG to other nodes.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmcommon.h |1 +
fs/ocfs2/dlm/dlmdomain.c |1 +
fs/ocfs2/dlm/dlmmaster.c | 33 +
3 files changed, 35 insertions
as leaving. Only block mles are expected
to be around, so we log ERROR for other mles but still free them.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmcommon.h |1 +
fs/ocfs2/dlm/dlmdomain.c |1 +
fs/ocfs2/dlm/dlmmaster.c | 34 ++
3
as leaving. Only block mles are expected
to be around, so we log ERROR for other mles but still free them.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmcommon.h |1 +
fs/ocfs2/dlm/dlmdomain.c |1 +
fs/ocfs2/dlm/dlmmaster.c | 40
## Block#TotalUsed Free Contig Size
052803072 322564294965350 34202182074032
..
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/suballoc.c | 16
1 files changed, 16 insertions(+), 0
## Block#TotalUsed Free Contig Size
052803072 322564294965350 34202182074032
..
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/suballoc.c | 16
1 files changed, 16 insertions(+), 0
From: Sunil Mushran sunil.mush...@oracle.com
Mainline fa16655a622e7c0fda76ca5155db6efc86968c65
A per-region debugfs file, elapsed_time_in_ms, shows the time since the
heartbeat timer was last armed.
Signed-off-by: Sunil Mushran sunil.mush...@oracle.com
---
fs/ocfs2/cluster/heartbeat.c | 46
The following two patches are backports from mainline to 1.4. These patches
create debugfs entry for heartbeat regions and to show elapsed time.
___
Ocfs2-devel mailing list
Ocfs2-devel@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-devel
From: Sunil Mushran sunil.mush...@oracle.com
Mainline 0841ed580fe8a3e51ba9dbb133dafc787cce428f
Signed-off-by: Sunil Mushran sunil.mush...@oracle.com
---
fs/ocfs2/cluster/heartbeat.c | 27 +++
1 files changed, 19 insertions(+), 8 deletions(-)
diff --git
I think I have seen this problem in ocfs2-1.2 and it was addressed by
using a new state DLM_LOCK_RES_IN_USE. But we didn't merge into mainline
as sunil suggested we need to look for a different approach
http://oss.oracle.com/pipermail/ocfs2-devel/2010-June/006669.html
+ spin_unlock(tmpres-spinlock);
+ spin_unlock(dlm-spinlock);
lockres could still get added to purgelist at this point and we
could still have the same problem? I think, here we need some
mechanism that marks the lockres is in use that would protect it
from adding to the
On 7/5/2011 11:17 PM, Sunil Mushran wrote:
2. All nodes have to scan all slots. Even live slots. I remember we
did for
a reason. And that reason should be in the comment in the patch written
by Srini.
When a node unlinks a file it inserts an entry into it's own orphan
slot. If another node
node 2's orphan dir. This
patch exists because that wasn't happening.
On 7/7/2011 1:26 PM, Sunil Mushran wrote:
On 07/07/2011 01:02 PM, Sunil Mushran wrote:
On 07/06/2011 11:19 PM, Srinivas Eeda wrote:
On 7/5/2011 11:17 PM, Sunil Mushran wrote:
2. All nodes have to scan all slots. Even live
When ocfs2dc thread holds dc_task_lock spinlock and receives soft IRQ for
I/O completion it deadlock itself trying to get same spinlock in
ocfs2_wake_downconvert_thread
The patch disables interrupts when acquiring dc_task_lock spinlock
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
Hi Tao,
thanks for reviewing.
When ocfs2dc thread holds dc_task_lock spinlock and receives soft IRQ for
I/O completion it deadlock itself trying to get same spinlock in
ocfs2_wake_downconvert_thread
could you please describe it in more detail?
When ocfs2dc thread is running on a cpu and
]
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlmglue.c | 30 ++
1 files changed, 18 insertions(+), 12 deletions(-)
diff --git a/fs/ocfs2/dlmglue.c b/fs/ocfs2/dlmglue.c
index 81a4cd2..d8552a5 100644
--- a/fs/ocfs2/dlmglue.c
+++ b/fs/ocfs2
]
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlmglue.c | 31 +++
1 files changed, 19 insertions(+), 12 deletions(-)
diff --git a/fs/ocfs2/dlmglue.c b/fs/ocfs2/dlmglue.c
index 81a4cd2..67af5db 100644
--- a/fs/ocfs2/dlmglue.c
+++ b/fs/ocfs2
sorry ignore this patch, resent another one after adding the new line.
On 1/30/2012 9:47 PM, Srinivas Eeda wrote:
When ocfs2dc thread holds dc_task_lock spinlock and receives soft IRQ it
deadlock itself trying to get same spinlock in ocfs2_wake_downconvert_thread.
Below is the stack snippet
[ocfs2_nodemanager]
[a030aac8] o2net_rx_until_empty+0x762/0x90d [ocfs2_nodemanager]
[81071802] worker_thread+0x14d/0x1ed
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmast.c |3 ++-
1 files changed, 2 insertions(+), 1 deletions(-)
diff --git a/fs/ocfs2/dlm
of
m_bitmap_ext_arr.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/reservations.c | 41 ++---
fs/ocfs2/reservations.h |7 ++-
2 files changed, 40 insertions(+), 8 deletions(-)
diff --git a/fs/ocfs2/reservations.c b/fs/ocfs2
chunks. In
best case there is only one record and increases as the filesystem gets
fragmented. Number of records at a time are limited depending on the size
of the bitmap and the max limit is defined by OCFS2_MAX_LOCAL_ALLOC_RECS.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2
Hi all,
can you please review following 3 patches that implement discontiguous
localalloc bitmap support for ocfs2 file system. This feature helps
applications that significantly fragment the filesystem.
These fixes needs changes to ocfs2 tools as well. I am sending those patches
for review
This patch adds supporting functions and modifies localalloc code to implement
discontiguous localalloc bitmap.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/localalloc.c | 523 -
1 files changed, 342 insertions(+), 181
Joel Becker wrote:
On Mon, May 07, 2012 at 04:21:27PM -0700, Srinivas Eeda wrote:
can you please review following 3 patches that implement discontiguous
localalloc bitmap support for ocfs2 file system. This feature helps
applications that significantly fragment the filesystem
Joel Becker wrote:
On Mon, May 07, 2012 at 04:21:29PM -0700, Srinivas Eeda wrote:
OH MY DOG NO. NEVER EVER DO THIS. You cannot update an old
filesystem on the fly! What about other nodes that are running older
versions of the software? They will crash or corrupt data
comments inline
On 5/24/2012 10:53 PM, xiaowei...@oracle.com wrote:
From: Xiaowei.Huxiaowei...@oracle.com
when the master requested locks ,but one/some of the live nodes died,
after it received the request msg and before send out the locks packages,
the recovery will fall into endless
Hi Jan,
thanks for helping.
Jan Kara wrote:
Hello,
his comments:
@ With those patches in, all other nodes will now queue downgrade of dentry
@ locks to ocfs2_wq thread. Then Node 1 gets a lock is in use when it calls
@ ocfs2_try_open_lock and so does other nodes and hence orphans lie
On 8/22/2012 7:17 AM, Jie Liu wrote:
Hi All,
These days, I am investigating an issue regarding OCFS2 unexpected
reboot in some real world use cases.
This problem occurred when the network status goes south, when the
disk IO load is too high, etc...
I suspect it might caused by ocfs2 fencing
Hi Joel, et al,
sorry for the delay in resending discontiguous localalloc patches. Can you
please review when you get a chance. I'll email the tools patches once kernel
patches are approved.
I came across two use cases where this feature will help.
1. On a customer site, after running an
will try to look for
atleast that big of chunk. If it's not available then it reduces the size by
half and retries.
#define OCFS2_DEFAULT_LOCALALLOC_CHUNK_SIZE (128*1024) /* 128kb */
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/localalloc.c | 28
of m_bitmap_ext_arr
Callers should pass number of discontiguous chunks during ocfs2_resmap_restart
and later call ocfs2_resmap_set_extent_size for every chunk.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/localalloc.c |4 +++-
fs/ocfs2/reservations.c | 39
and as the filesystem gets
fragmented you may see multipe records.
This feature can be enabled/disabled when the file system is offline.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/localalloc.c | 22 ++--
fs/ocfs2/ocfs2.h |7 +
fs/ocfs2/ocfs2_fs.h
This patch adds code to support discontiguous localalloc bitmap. At any given
time there can be a combination of volumes that have discontigous feature
enabled or disabled.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/localalloc.c | 478
ocfs2_local_alloc_inode is used in multiple functions. It is convenient if we
move ocfs2_local_alloc_inode to ocfs2 super.
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/localalloc.c | 53 -
fs/ocfs2/ocfs2.h |1 +
2
On 05/31/2013 03:38 AM, Xue jiufei wrote:
Hi, Xiaowei
It's OK to simlify the patch just as you did. But we don't want to resend
master request to all others nodes in consideration of network traffic.
So we record those maybe down nodes in down_nodemap.
δΊ 2013/5/28 14:12, xiaowei.hu ει:
Hi,
is unset by migrate handler thread. umount
thread will ultimately migrate this lockres during another pass of the
lockres hash list.
Signed-off-by: Tariq Saeed tariq.x.sa...@oracle.com
Signed-off-by: Srinivas Eeda srinivas.e...@oracle.com
---
fs/ocfs2/dlm/dlmmaster.c |4
1 files
Hi Goldwyn,
On 01/08/2014 04:12 PM, Goldwyn Rodrigues wrote:
Hi,
From the comments in fs/ocfs2/inode.h:90 it seems, this was used in
legacy ocfs2 systems when a node received unlink votes. Since unlink
votes has been done away with and replaced with open locks, is this
flag still required?
1 - 100 of 146 matches
Mail list logo