Commit 494e7fe591bf for kernel

commit 494e7fe591bf834d57c6607cdc26ab8873708aa7
Merge: fa593d0f969d 6ffb9017e932
Author: Linus Torvalds <torvalds@linux-foundation.org>
Date:   Sun Mar 30 13:06:27 2025 -0700

    Merge tag 'bpf_res_spin_lock' of git://git.kernel.org/pub/scm/linux/kernel/git/bpf/bpf-next

    Pull bpf relisient spinlock support from Alexei Starovoitov:
     "This patch set introduces Resilient Queued Spin Lock (or rqspinlock
      with res_spin_lock() and res_spin_unlock() APIs).

      This is a qspinlock variant which recovers the kernel from a stalled
      state when the lock acquisition path cannot make forward progress.
      This can occur when a lock acquisition attempt enters a deadlock
      situation (e.g. AA, or ABBA), or more generally, when the owner of the
      lock (which we’re trying to acquire) isn’t making forward progress.
      Deadlock detection is the main mechanism used to provide instant
      recovery, with the timeout mechanism acting as a final line of
      defense. Detection is triggered immediately when beginning the waiting
      loop of a lock slow path.

      Additionally, BPF programs attached to different parts of the kernel
      can introduce new control flow into the kernel, which increases the
      likelihood of deadlocks in code not written to handle reentrancy.
      There have been multiple syzbot reports surfacing deadlocks in
      internal kernel code due to the diverse ways in which BPF programs can
      be attached to different parts of the kernel. By switching the BPF
      subsystem’s lock usage to rqspinlock, all of these issues are
      mitigated at runtime.

      This spin lock implementation allows BPF maps to become safer and
      remove mechanisms that have fallen short in assuring safety when
      nesting programs in arbitrary ways in the same context or across
      different contexts.

      We run benchmarks that stress locking scalability and perform
      comparison against the baseline (qspinlock). For the rqspinlock case,
      we replace the default qspinlock with it in the kernel, such that all
      spin locks in the kernel use the rqspinlock slow path. As such,
      benchmarks that stress kernel spin locks end up exercising rqspinlock.

      More details in the cover letter in commit 6ffb9017e932 ("Merge branch
      'resilient-queued-spin-lock'")"

    * tag 'bpf_res_spin_lock' of git://git.kernel.org/pub/scm/linux/kernel/git/bpf/bpf-next: (24 commits)
      selftests/bpf: Add tests for rqspinlock
      bpf: Maintain FIFO property for rqspinlock unlock
      bpf: Implement verifier support for rqspinlock
      bpf: Introduce rqspinlock kfuncs
      bpf: Convert lpm_trie.c to rqspinlock
      bpf: Convert percpu_freelist.c to rqspinlock
      bpf: Convert hashtab.c to rqspinlock
      rqspinlock: Add locktorture support
      rqspinlock: Add entry to Makefile, MAINTAINERS
      rqspinlock: Add macros for rqspinlock usage
      rqspinlock: Add basic support for CONFIG_PARAVIRT
      rqspinlock: Add a test-and-set fallback
      rqspinlock: Add deadlock detection and recovery
      rqspinlock: Protect waiters in trylock fallback from stalls
      rqspinlock: Protect waiters in queue from stalls
      rqspinlock: Protect pending bit owners from stalls
      rqspinlock: Hardcode cond_acquire loops for arm64
      rqspinlock: Add support for timeouts
      rqspinlock: Drop PV and virtualization support
      rqspinlock: Add rqspinlock.h header
      ...