]> git.ipfire.org Git - thirdparty/glibc.git/commitdiff
Add LLL_MUTEX_READ_LOCK [BZ #28537]
authorH.J. Lu <hjl.tools@gmail.com>
Wed, 3 Nov 2021 01:33:07 +0000 (18:33 -0700)
committerSunil K Pandey <skpgkp2@gmail.com>
Wed, 28 Sep 2022 20:47:29 +0000 (13:47 -0700)
CAS instruction is expensive.  From the x86 CPU's point of view, getting
a cache line for writing is more expensive than reading.  See Appendix
A.2 Spinlock in:

https://www.intel.com/content/dam/www/public/us/en/documents/white-papers/xeon-lock-scaling-analysis-paper.pdf

The full compare and swap will grab the cache line exclusive and cause
excessive cache line bouncing.

Add LLL_MUTEX_READ_LOCK to do an atomic load and skip CAS in spinlock
loop if compare may fail to reduce cache line bouncing on contended locks.

Reviewed-by: Szabolcs Nagy <szabolcs.nagy@arm.com>
(cherry picked from commit d672a98a1af106bd68deb15576710cd61363f7a6)

nptl/pthread_mutex_lock.c

index 0fd7467b97680bcd271d32fa529567205b4f0600..ac8a472c08d6d0b84726204c47d027fece73eaff 100644 (file)
 #define FORCE_ELISION(m, s)
 #endif
 
+#ifndef LLL_MUTEX_READ_LOCK
+# define LLL_MUTEX_READ_LOCK(mutex) \
+  atomic_load_relaxed (&(mutex)->__data.__lock)
+#endif
+
 static int __pthread_mutex_lock_full (pthread_mutex_t *mutex)
      __attribute_noinline__;
 
@@ -133,6 +138,8 @@ __pthread_mutex_lock (pthread_mutex_t *mutex)
                  break;
                }
              atomic_spin_nop ();
+             if (LLL_MUTEX_READ_LOCK (mutex) != 0)
+               continue;
            }
          while (LLL_MUTEX_TRYLOCK (mutex) != 0);