public inbox for lvm2-cvs@sourceware.org
help / color / mirror / Atom feed
From: mbroz@sourceware.org
To: lvm-devel@redhat.com, lvm2-cvs@sourceware.org
Subject: LVM2 ./WHATS_NEW daemons/clvmd/lvm-functions.c
Date: Wed, 09 Dec 2009 19:00:00 -0000	[thread overview]
Message-ID: <20091209190017.22077.qmail@sourceware.org> (raw)

CVSROOT:	/cvs/lvm2
Module name:	LVM2
Changes by:	mbroz@sourceware.org	2009-12-09 19:00:16

Modified files:
	.              : WHATS_NEW 
	daemons/clvmd  : lvm-functions.c 

Log message:
	Never ever use distributed lock for LV in non-clustered VG.
	
	The LV locks make sense only for clustered LVs.
	
	Properly check cluster flag and never issue cluster lock here.
	
	There are several places in code, where it is already checked, this
	patch add this check to all needed calls.
	
	In previous code the lock behaviour was inconsistent,
	for example, the pre/post callback can take lock even for local volume,
	but deactivate call do not released this lock and it remains held forever.
	
	The local LV lock request now just let run the underlying activation code
	on local node, the same process like in local locking.
	
	(Again, this is important for new mirror repair calls, here for local
	mirrors but with cluster locking enabled.)

Patches:
http://sourceware.org/cgi-bin/cvsweb.cgi/LVM2/WHATS_NEW.diff?cvsroot=lvm2&r1=1.1346&r2=1.1347
http://sourceware.org/cgi-bin/cvsweb.cgi/LVM2/daemons/clvmd/lvm-functions.c.diff?cvsroot=lvm2&r1=1.75&r2=1.76

--- LVM2/WHATS_NEW	2009/12/09 18:55:53	1.1346
+++ LVM2/WHATS_NEW	2009/12/09 19:00:16	1.1347
@@ -1,5 +1,6 @@
 Version 2.02.57 -
 ====================================
+  Never use distributed lock for LV in non-clustered VG.
   Allow implicit lock conversion in pre/post callbacks on local node.
   Allow implicit convert to the same cluster lock mode.
   Replace magic masks in cluster locking code by defined masks.
--- LVM2/daemons/clvmd/lvm-functions.c	2009/12/09 18:55:53	1.75
+++ LVM2/daemons/clvmd/lvm-functions.c	2009/12/09 19:00:16	1.76
@@ -331,7 +331,8 @@
 
 	/* Is it already open ? */
 	oldmode = get_current_lock(resource);
-	if (oldmode == mode) {
+	if (oldmode == mode && (lock_flags & LCK_CLUSTER_VG)) {
+		DEBUGLOG("do_activate_lv, lock already held at %d\n", oldmode);
 		return 0;	/* Nothing to do */
 	}
 
@@ -384,13 +385,13 @@
 }
 
 /* Resume the LV if it was active */
-static int do_resume_lv(char *resource)
+static int do_resume_lv(char *resource, unsigned char lock_flags)
 {
 	int oldmode;
 
 	/* Is it open ? */
 	oldmode = get_current_lock(resource);
-	if (oldmode == -1) {
+	if (oldmode == -1 && (lock_flags & LCK_CLUSTER_VG)) {
 		DEBUGLOG("do_resume_lv, lock not already held\n");
 		return 0;	/* We don't need to do anything */
 	}
@@ -402,15 +403,15 @@
 }
 
 /* Suspend the device if active */
-static int do_suspend_lv(char *resource)
+static int do_suspend_lv(char *resource, unsigned char lock_flags)
 {
 	int oldmode;
 	struct lvinfo lvi;
 
 	/* Is it open ? */
 	oldmode = get_current_lock(resource);
-	if (oldmode == -1) {
-		DEBUGLOG("do_suspend_lv, lock held at %d\n", oldmode);
+	if (oldmode == -1 && (lock_flags & LCK_CLUSTER_VG)) {
+		DEBUGLOG("do_suspend_lv, lock not already held\n");
 		return 0; /* Not active, so it's OK */
 	}
 
@@ -502,14 +503,14 @@
 		break;
 
 	case LCK_LV_SUSPEND:
-		status = do_suspend_lv(resource);
+		status = do_suspend_lv(resource, lock_flags);
 		if (!status)
 			suspended++;
 		break;
 
 	case LCK_UNLOCK:
 	case LCK_LV_RESUME:	/* if active */
-		status = do_resume_lv(resource);
+		status = do_resume_lv(resource, lock_flags);
 		if (!status)
 			suspended--;
 		break;
@@ -552,7 +553,8 @@
 	   before suspending cluster-wide.
 	   LKF_CONVERT is used always, local node is going to modify metadata
 	 */
-	if ((command & (LCK_SCOPE_MASK | LCK_TYPE_MASK)) == LCK_LV_SUSPEND) {
+	if ((command & (LCK_SCOPE_MASK | LCK_TYPE_MASK)) == LCK_LV_SUSPEND &&
+	    (lock_flags & LCK_CLUSTER_VG)) {
 		DEBUGLOG("pre_lock_lv: resource '%s', cmd = %s, flags = %s\n",
 			 resource, decode_locking_cmd(command), decode_flags(lock_flags));
 
@@ -569,7 +571,8 @@
 	int status;
 
 	/* Opposite of above, done on resume after a metadata update */
-	if ((command & (LCK_SCOPE_MASK | LCK_TYPE_MASK)) == LCK_LV_RESUME) {
+	if ((command & (LCK_SCOPE_MASK | LCK_TYPE_MASK)) == LCK_LV_RESUME &&
+	    (lock_flags & LCK_CLUSTER_VG)) {
 		int oldmode;
 
 		DEBUGLOG


             reply	other threads:[~2009-12-09 19:00 UTC|newest]

Thread overview: 23+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2009-12-09 19:00 mbroz [this message]
  -- strict thread matches above, loose matches on Subject: below --
2012-03-01  9:54 zkabelac
2012-02-23 22:50 zkabelac
2011-08-10 11:00 mbroz
2011-03-02 23:38 mbroz
2010-07-01 21:23 mbroz
2010-03-24 22:25 snitzer
2010-01-26  8:00 mbroz
2009-12-09 19:01 mbroz
2009-12-09 18:55 mbroz
2009-12-09 18:45 mbroz
2009-10-22 17:45 agk
2009-10-01 14:14 ccaulfield
2009-08-05 14:18 ccaulfield
2009-03-05 16:25 mbroz
2008-04-08 13:03 ccaulfield
2008-04-04  8:53 ccaulfield
2007-08-24  8:29 pcaulfield
2007-08-23 12:19 pcaulfield
2007-04-24 15:13 pcaulfield
2005-02-14  9:07 pcaulfield
2005-02-02 11:42 pcaulfield
2004-10-04  9:23 pcaulfield

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20091209190017.22077.qmail@sourceware.org \
    --to=mbroz@sourceware.org \
    --cc=lvm-devel@redhat.com \
    --cc=lvm2-cvs@sourceware.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).