| 24811637dbfd ("locking/lock_events: Use raw_cpu_{add,inc}() for stats") |
| 51816e9e1139 ("locking/lock_events: Use this_cpu_add() when necessary") |
| ad53fa10fa9e ("locking/qspinlock_stat: Introduce generic lockevent_*() counting APIs") |
| 412f34a82ccf ("locking/qspinlock_stat: Track the no MCS node available case") |
| d682b596d993 ("locking/qspinlock: Handle > 4 slowpath nesting levels") |
| 0fa809ca7f81 ("locking/pvqspinlock: Extend node size when pvqspinlock is configured") |
| 1222109a5363 ("locking/qspinlock_stat: Count instances of nested lock slowpaths") |
| 53bf57fab732 ("locking/qspinlock: Re-order code") |
| 81d3dc9a349b ("locking/qspinlock: Add stat tracking for pending vs. slowpath") |
| f9c811fac48c ("locking/qspinlock: Use atomic_cond_read_acquire()") |
| c61da58d8a9b ("locking/qspinlock: Kill cmpxchg() loop when claiming lock from head of queue") |
| 59fb586b4a07 ("locking/qspinlock: Remove unbounded cmpxchg() loop from locking slowpath") |
| 625e88be1f41 ("locking/qspinlock: Merge 'struct __qspinlock' into 'struct qspinlock'") |
| 11dc13224c97 ("locking/qspinlock: Ensure node->count is updated before initialising node") |
| 11752adb68a3 ("locking/pvqspinlock: Implement hybrid PV queued/unfair locks") |