]> git.ipfire.org Git - thirdparty/postgresql.git/commitdiff
Fix segfault from releasing locks in detached DSM segments
authorAmit Langote <amitlan@postgresql.org>
Fri, 16 Jan 2026 04:01:52 +0000 (13:01 +0900)
committerAmit Langote <amitlan@postgresql.org>
Fri, 16 Jan 2026 04:14:21 +0000 (13:14 +0900)
If a FATAL error occurs while holding a lock in a DSM segment (such
as a dshash lock) and the process is not in a transaction, a
segmentation fault can occur during process exit.

The problem sequence is:

 1. Process acquires a lock in a DSM segment (e.g., via dshash)
 2. FATAL error occurs outside transaction context
 3. proc_exit() begins, calling before_shmem_exit callbacks
 4. dsm_backend_shutdown() detaches all DSM segments
 5. Later, on_shmem_exit callbacks run
 6. ProcKill() calls LWLockReleaseAll()
 7. Segfault: the lock being released is in unmapped memory

This only manifests outside transaction contexts because
AbortTransaction() calls LWLockReleaseAll() during transaction
abort, releasing locks before DSM cleanup. Background workers and
other non-transactional code paths are vulnerable.

Fix by calling LWLockReleaseAll() unconditionally at the start of
shmem_exit(), before any callbacks run. Releasing locks before
callbacks prevents the segfault - locks must be released before
dsm_backend_shutdown() detaches their memory. This is safe because
after an error, held locks are protecting potentially inconsistent
data anyway, and callbacks can acquire fresh locks if needed.

Also add a comment noting that LWLockReleaseAll() must be safe to
call before LWLock initialization (which it is, since
num_held_lwlocks will be 0), plus an Assert for the post-condition.

This fix aligns with the original design intent from commit
001a573a2, which noted that backends must clean up shared memory
state (including releasing lwlocks) before unmapping dynamic shared
memory segments.

Reported-by: Rahila Syed <rahilasyed90@gmail.com>
Author: Rahila Syed <rahilasyed90@gmail.com>
Reviewed-by: Amit Langote <amitlangote09@gmail.com>
Reviewed-by: Dilip Kumar <dilipbalaut@gmail.com>
Reviewed-by: Andres Freund <andres@anarazel.de>
Reviewed-by: Álvaro Herrera <alvherre@kurilemu.de>
Discussion: https://postgr.es/m/CAH2L28uSvyiosL+kaic9249jRVoQiQF6JOnaCitKFq=xiFzX3g@mail.gmail.com
Backpatch-through: 14

src/backend/storage/ipc/ipc.c
src/backend/storage/lmgr/lwlock.c

index 567739b5be93a33404d4dc70acac2b500c28c9f4..0684d3434688b79f5354dd0d8c493b14eec5ccd7 100644 (file)
@@ -29,6 +29,7 @@
 #endif
 #include "storage/dsm.h"
 #include "storage/ipc.h"
+#include "storage/lwlock.h"
 #include "tcop/tcopprot.h"
 
 
@@ -229,13 +230,19 @@ shmem_exit(int code)
 {
        shmem_exit_inprogress = true;
 
+       /*
+        * Release any LWLocks we might be holding before callbacks run. This
+        * prevents accessing locks in detached DSM segments and allows callbacks
+        * to acquire new locks.
+        */
+       LWLockReleaseAll();
+
        /*
         * Call before_shmem_exit callbacks.
         *
         * These should be things that need most of the system to still be up and
         * working, such as cleanup of temp relations, which requires catalog
-        * access; or things that need to be completed because later cleanup steps
-        * depend on them, such as releasing lwlocks.
+        * access.
         */
        elog(DEBUG3, "shmem_exit(%d): %d before_shmem_exit callbacks to make",
                 code, before_shmem_exit_index);
index fc4b42ea5d6aa573ac6d30d5d871d39b6ca8d4e2..c3d4b7275ecc14c83c64b3a4a430a9de542905a0 100644 (file)
@@ -1946,6 +1946,10 @@ LWLockReleaseClearVar(LWLock *lock, pg_atomic_uint64 *valptr, uint64 val)
  * unchanged by this operation.  This is necessary since InterruptHoldoffCount
  * has been set to an appropriate level earlier in error recovery. We could
  * decrement it below zero if we allow it to drop for each released lock!
+ *
+ * Note that this function must be safe to call even before the LWLock
+ * subsystem has been initialized (e.g., during early startup failures).
+ * In that case, num_held_lwlocks will be 0 and we do nothing.
  */
 void
 LWLockReleaseAll(void)
@@ -1956,6 +1960,8 @@ LWLockReleaseAll(void)
 
                LWLockRelease(held_lwlocks[num_held_lwlocks - 1].lock);
        }
+
+       Assert(num_held_lwlocks == 0);
 }