]> granicus.if.org Git - postgresql/commitdiff
Revert "Avoid spurious deadlocks when upgrading a tuple lock"
authorAlvaro Herrera <alvherre@alvh.no-ip.org>
Mon, 17 Jun 2019 02:24:20 +0000 (22:24 -0400)
committerAlvaro Herrera <alvherre@alvh.no-ip.org>
Mon, 17 Jun 2019 02:24:20 +0000 (22:24 -0400)
This reverts commits 3da73d6839dc and de87a084c0a5.

This code has some tricky corner cases that I'm not sure are correct and
not properly tested anyway, so I'm reverting the whole thing for next
week's releases (reintroducing the deadlock bug that we set to fix).
I'll try again afterwards.

Discussion: https://postgr.es/m/E1hbXKQ-0003g1-0C@gemulon.postgresql.org

src/backend/access/heap/README.tuplock
src/backend/access/heap/heapam.c
src/test/isolation/expected/tuplelock-upgrade-no-deadlock.out [deleted file]
src/test/isolation/isolation_schedule
src/test/isolation/specs/tuplelock-upgrade-no-deadlock.spec [deleted file]

index 9674d0c0f0706d67d28501654af3206ef9d14d1c..10b8d78ab7e0d40964698d8e6bf66614f96cb214 100644 (file)
@@ -36,16 +36,6 @@ do LockTuple as well, if there is any conflict, to ensure that they don't
 starve out waiting exclusive-lockers.  However, if there is not any active
 conflict for a tuple, we don't incur any extra overhead.
 
-We make an exception to the above rule for those lockers that already hold
-some lock on a tuple and attempt to acquire a stronger one on it.  In that
-case, we skip the LockTuple() call even when there are conflicts, provided
-that the target tuple is being locked, updated or deleted by multiple sessions
-concurrently.  Failing to skip the lock would risk a deadlock, e.g., between a
-session that was first to record its weaker lock in the tuple header and would
-be waiting on the LockTuple() call to upgrade to the stronger lock level, and
-another session that has already done LockTuple() and is waiting for the first
-session transaction to release its tuple header-level lock.
-
 We provide four levels of tuple locking strength: SELECT FOR UPDATE obtains an
 exclusive lock which prevents any kind of modification of the tuple. This is
 the lock level that is implicitly taken by DELETE operations, and also by
index 6627a36aaf2addfc16b715d9d0d98ed08862e37b..8fecf7113e62e9bd96ad8ccd838d9fb71c94530f 100644 (file)
@@ -117,7 +117,7 @@ static void GetMultiXactIdHintBits(MultiXactId multi, uint16 *new_infomask,
 static TransactionId MultiXactIdGetUpdateXid(TransactionId xmax,
                                                uint16 t_infomask);
 static bool DoesMultiXactIdConflict(MultiXactId multi, uint16 infomask,
-                                               LockTupleMode lockmode, bool *current_is_member);
+                                               LockTupleMode lockmode);
 static void MultiXactIdWait(MultiXactId multi, MultiXactStatus status, uint16 infomask,
                                Relation rel, ItemPointer ctid, XLTW_Oper oper,
                                int *remaining);
@@ -3087,20 +3087,15 @@ l1:
                 */
                if (infomask & HEAP_XMAX_IS_MULTI)
                {
-                       bool            current_is_member = false;
-
+                       /* wait for multixact */
                        if (DoesMultiXactIdConflict((MultiXactId) xwait, infomask,
-                                                                               LockTupleExclusive, &current_is_member))
+                                                                               LockTupleExclusive))
                        {
                                LockBuffer(buffer, BUFFER_LOCK_UNLOCK);
 
-                               /*
-                                * Acquire the lock, if necessary (but skip it when we're
-                                * requesting a lock and already have one; avoids deadlock).
-                                */
-                               if (!current_is_member)
-                                       heap_acquire_tuplock(relation, &(tp.t_self), LockTupleExclusive,
-                                                                                LockWaitBlock, &have_tuple_lock);
+                               /* acquire tuple lock, if necessary */
+                               heap_acquire_tuplock(relation, &(tp.t_self), LockTupleExclusive,
+                                                                        LockWaitBlock, &have_tuple_lock);
 
                                /* wait for multixact */
                                MultiXactIdWait((MultiXactId) xwait, MultiXactStatusUpdate, infomask,
@@ -3664,20 +3659,15 @@ l2:
                {
                        TransactionId update_xact;
                        int                     remain;
-                       bool            current_is_member = false;
 
                        if (DoesMultiXactIdConflict((MultiXactId) xwait, infomask,
-                                                                               *lockmode, &current_is_member))
+                                                                               *lockmode))
                        {
                                LockBuffer(buffer, BUFFER_LOCK_UNLOCK);
 
-                               /*
-                                * Acquire the lock, if necessary (but skip it when we're
-                                * requesting a lock and already have one; avoids deadlock).
-                                */
-                               if (!current_is_member)
-                                       heap_acquire_tuplock(relation, &(oldtup.t_self), *lockmode,
-                                                                                LockWaitBlock, &have_tuple_lock);
+                               /* acquire tuple lock, if necessary */
+                               heap_acquire_tuplock(relation, &(oldtup.t_self), *lockmode,
+                                                                        LockWaitBlock, &have_tuple_lock);
 
                                /* wait for multixact */
                                MultiXactIdWait((MultiXactId) xwait, mxact_status, infomask,
@@ -4636,7 +4626,6 @@ l3:
                uint16          infomask;
                uint16          infomask2;
                bool            require_sleep;
-               bool            skip_tuple_lock = false;
                ItemPointerData t_ctid;
 
                /* must copy state data before unlocking buffer */
@@ -4690,21 +4679,6 @@ l3:
                                                result = HeapTupleMayBeUpdated;
                                                goto out_unlocked;
                                        }
-                                       else
-                                       {
-                                               /*
-                                                * Disable acquisition of the heavyweight tuple lock.
-                                                * Otherwise, when promoting a weaker lock, we might
-                                                * deadlock with another locker that has acquired the
-                                                * heavyweight tuple lock and is waiting for our
-                                                * transaction to finish.
-                                                *
-                                                * Note that in this case we still need to wait for
-                                                * the multixact if required, to avoid acquiring
-                                                * conflicting locks.
-                                                */
-                                               skip_tuple_lock = true;
-                                       }
                                }
 
                                if (members)
@@ -4859,7 +4833,7 @@ l3:
                        if (infomask & HEAP_XMAX_IS_MULTI)
                        {
                                if (!DoesMultiXactIdConflict((MultiXactId) xwait, infomask,
-                                                                                        mode, NULL))
+                                                                                        mode))
                                {
                                        /*
                                         * No conflict, but if the xmax changed under us in the
@@ -4936,15 +4910,13 @@ l3:
                        /*
                         * Acquire tuple lock to establish our priority for the tuple, or
                         * die trying.  LockTuple will release us when we are next-in-line
-                        * for the tuple.  We must do this even if we are share-locking,
-                        * but not if we already have a weaker lock on the tuple.
+                        * for the tuple.  We must do this even if we are share-locking.
                         *
                         * If we are forced to "start over" below, we keep the tuple lock;
                         * this arranges that we stay at the head of the line while
                         * rechecking tuple state.
                         */
-                       if (!skip_tuple_lock &&
-                               !heap_acquire_tuplock(relation, tid, mode, wait_policy,
+                       if (!heap_acquire_tuplock(relation, tid, mode, wait_policy,
                                                                          &have_tuple_lock))
                        {
                                /*
@@ -7116,13 +7088,10 @@ HeapTupleGetUpdateXid(HeapTupleHeader tuple)
  * tuple lock of the given strength?
  *
  * The passed infomask pairs up with the given multixact in the tuple header.
- *
- * If current_is_member is not NULL, it is set to 'true' if the current
- * transaction is a member of the given multixact.
  */
 static bool
 DoesMultiXactIdConflict(MultiXactId multi, uint16 infomask,
-                                               LockTupleMode lockmode, bool *current_is_member)
+                                               LockTupleMode lockmode)
 {
        int                     nmembers;
        MultiXactMember *members;
@@ -7143,26 +7112,17 @@ DoesMultiXactIdConflict(MultiXactId multi, uint16 infomask,
                        TransactionId memxid;
                        LOCKMODE        memlockmode;
 
-                       if (result && (current_is_member == NULL || *current_is_member))
-                               break;
-
                        memlockmode = LOCKMODE_from_mxstatus(members[i].status);
 
-                       /* ignore members from current xact (but track their presence) */
-                       memxid = members[i].xid;
-                       if (TransactionIdIsCurrentTransactionId(memxid))
-                       {
-                               if (current_is_member != NULL)
-                                       *current_is_member = true;
-                               continue;
-                       }
-                       else if (result)
-                               continue;
-
                        /* ignore members that don't conflict with the lock we want */
                        if (!DoLockModesConflict(memlockmode, wanted))
                                continue;
 
+                       /* ignore members from current xact */
+                       memxid = members[i].xid;
+                       if (TransactionIdIsCurrentTransactionId(memxid))
+                               continue;
+
                        if (ISUPDATE_from_mxstatus(members[i].status))
                        {
                                /* ignore aborted updaters */
@@ -7179,11 +7139,10 @@ DoesMultiXactIdConflict(MultiXactId multi, uint16 infomask,
                        /*
                         * Whatever remains are either live lockers that conflict with our
                         * wanted lock, and updaters that are not aborted.  Those conflict
-                        * with what we want.  Set up to return true, but keep going to
-                        * look for the current transaction among the multixact members,
-                        * if needed.
+                        * with what we want, so return true.
                         */
                        result = true;
+                       break;
                }
                pfree(members);
        }
diff --git a/src/test/isolation/expected/tuplelock-upgrade-no-deadlock.out b/src/test/isolation/expected/tuplelock-upgrade-no-deadlock.out
deleted file mode 100644 (file)
index 5148aa2..0000000
+++ /dev/null
@@ -1,150 +0,0 @@
-Parsed test spec with 3 sessions
-
-starting permutation: s1_share s2_for_update s3_share s3_for_update s1_rollback s3_rollback s2_rollback
-step s1_share: select id from tlu_job where id = 1 for share;
-id             
-
-1              
-step s2_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s3_share: select id from tlu_job where id = 1 for share;
-id             
-
-1              
-step s3_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s1_rollback: rollback;
-step s3_for_update: <... completed>
-id             
-
-1              
-step s3_rollback: rollback;
-step s2_for_update: <... completed>
-id             
-
-1              
-step s2_rollback: rollback;
-
-starting permutation: s1_keyshare s2_for_update s3_keyshare s1_update s3_update s1_rollback s3_rollback s2_rollback
-step s1_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s2_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s3_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s1_update: update tlu_job set name = 'b' where id = 1;
-step s3_update: update tlu_job set name = 'c' where id = 1; <waiting ...>
-step s1_rollback: rollback;
-step s3_update: <... completed>
-step s3_rollback: rollback;
-step s2_for_update: <... completed>
-id             
-
-1              
-step s2_rollback: rollback;
-
-starting permutation: s1_keyshare s2_for_update s3_keyshare s1_update s3_update s1_commit s3_rollback s2_rollback
-step s1_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s2_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s3_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s1_update: update tlu_job set name = 'b' where id = 1;
-step s3_update: update tlu_job set name = 'c' where id = 1; <waiting ...>
-step s1_commit: commit;
-step s3_update: <... completed>
-step s3_rollback: rollback;
-step s2_for_update: <... completed>
-id             
-
-1              
-step s2_rollback: rollback;
-
-starting permutation: s1_keyshare s2_for_update s3_keyshare s3_delete s1_rollback s3_rollback s2_rollback
-step s1_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s2_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s3_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s3_delete: delete from tlu_job where id = 1; <waiting ...>
-step s1_rollback: rollback;
-step s3_delete: <... completed>
-step s3_rollback: rollback;
-step s2_for_update: <... completed>
-id             
-
-1              
-step s2_rollback: rollback;
-
-starting permutation: s1_keyshare s2_for_update s3_keyshare s3_delete s1_rollback s3_commit s2_rollback
-step s1_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s2_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s3_keyshare: select id from tlu_job where id = 1 for key share;
-id             
-
-1              
-step s3_delete: delete from tlu_job where id = 1; <waiting ...>
-step s1_rollback: rollback;
-step s3_delete: <... completed>
-step s3_commit: commit;
-step s2_for_update: <... completed>
-id             
-
-step s2_rollback: rollback;
-
-starting permutation: s1_share s2_for_update s3_for_update s1_rollback s2_rollback s3_rollback
-step s1_share: select id from tlu_job where id = 1 for share;
-id             
-
-1              
-step s2_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s3_for_update: select id from tlu_job where id = 1 for update; <waiting ...>
-step s1_rollback: rollback;
-step s2_for_update: <... completed>
-id             
-
-1              
-step s2_rollback: rollback;
-step s3_for_update: <... completed>
-id             
-
-1              
-step s3_rollback: rollback;
-
-starting permutation: s1_share s2_update s3_update s1_rollback s2_rollback s3_rollback
-step s1_share: select id from tlu_job where id = 1 for share;
-id             
-
-1              
-step s2_update: update tlu_job set name = 'b' where id = 1; <waiting ...>
-step s3_update: update tlu_job set name = 'c' where id = 1; <waiting ...>
-step s1_rollback: rollback;
-step s2_update: <... completed>
-step s2_rollback: rollback;
-step s3_update: <... completed>
-step s3_rollback: rollback;
-
-starting permutation: s1_share s2_delete s3_delete s1_rollback s2_rollback s3_rollback
-step s1_share: select id from tlu_job where id = 1 for share;
-id             
-
-1              
-step s2_delete: delete from tlu_job where id = 1; <waiting ...>
-step s3_delete: delete from tlu_job where id = 1; <waiting ...>
-step s1_rollback: rollback;
-step s2_delete: <... completed>
-step s2_rollback: rollback;
-step s3_delete: <... completed>
-step s3_rollback: rollback;
index a9efc40709be99d62cbdc6ddb764689283fc1946..bc8198890ba7ea7bade02ba73396cba7d853fcec 100644 (file)
@@ -43,7 +43,6 @@ test: update-locked-tuple
 test: propagate-lock-delete
 test: tuplelock-conflict
 test: tuplelock-update
-test: tuplelock-upgrade-no-deadlock
 test: freeze-the-dead
 test: nowait
 test: nowait-2
diff --git a/src/test/isolation/specs/tuplelock-upgrade-no-deadlock.spec b/src/test/isolation/specs/tuplelock-upgrade-no-deadlock.spec
deleted file mode 100644 (file)
index 431fd6f..0000000
+++ /dev/null
@@ -1,57 +0,0 @@
-# This test checks that multiple sessions locking a single row in a table
-# do not deadlock each other when one of them upgrades its existing lock
-# while the others are waiting for it.
-
-setup
-{
-    drop table if exists tlu_job;
-    create table tlu_job (id integer primary key, name text);
-
-    insert into tlu_job values(1, 'a');
-}
-
-
-teardown
-{
-    drop table tlu_job;
-}
-
-session "s1"
-setup { begin; }
-step "s1_keyshare" { select id from tlu_job where id = 1 for key share;}
-step "s1_share" { select id from tlu_job where id = 1 for share; }
-step "s1_update" { update tlu_job set name = 'b' where id = 1;  }
-step "s1_delete" { delete from tlu_job where id = 1; }
-step "s1_rollback" { rollback; }
-step "s1_commit" { commit; }
-
-session "s2"
-setup { begin; }
-step "s2_for_update" { select id from tlu_job where id = 1 for update; }
-step "s2_update" { update tlu_job set name = 'b' where id = 1; }
-step "s2_delete" { delete from tlu_job where id = 1; }
-step "s2_rollback" { rollback; }
-step "s2_commit" { commit; }
-
-session "s3"
-setup { begin; }
-step "s3_keyshare" { select id from tlu_job where id = 1 for key share; }
-step "s3_share" { select id from tlu_job where id = 1 for share; }
-step "s3_for_update" { select id from tlu_job where id = 1 for update; }
-step "s3_update" { update tlu_job set name = 'c' where id = 1; }
-step "s3_delete" { delete from tlu_job where id = 1; }
-step "s3_rollback" { rollback; }
-step "s3_commit" { commit; }
-
-# test that s2 will not deadlock with s3 when s1 is rolled back
-permutation "s1_share" "s2_for_update" "s3_share" "s3_for_update" "s1_rollback" "s3_rollback" "s2_rollback"
-# test that update does not cause deadlocks if it can proceed
-permutation  "s1_keyshare" "s2_for_update" "s3_keyshare" "s1_update" "s3_update" "s1_rollback" "s3_rollback" "s2_rollback"
-permutation  "s1_keyshare" "s2_for_update" "s3_keyshare" "s1_update" "s3_update" "s1_commit"   "s3_rollback" "s2_rollback"
-# test that delete does not cause deadlocks if it can proceed
-permutation "s1_keyshare" "s2_for_update" "s3_keyshare"  "s3_delete" "s1_rollback" "s3_rollback" "s2_rollback"
-permutation "s1_keyshare" "s2_for_update" "s3_keyshare"  "s3_delete" "s1_rollback" "s3_commit"   "s2_rollback"
-# test that sessions that don't upgrade locks acquire them in order
-permutation "s1_share" "s2_for_update" "s3_for_update" "s1_rollback" "s2_rollback" "s3_rollback"
-permutation "s1_share" "s2_update" "s3_update" "s1_rollback" "s2_rollback" "s3_rollback"
-permutation "s1_share" "s2_delete" "s3_delete" "s1_rollback" "s2_rollback" "s3_rollback"