smp_lock.h 1.4 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152
  1. #ifndef __LINUX_SMPLOCK_H
  2. #define __LINUX_SMPLOCK_H
  3. #ifdef CONFIG_LOCK_KERNEL
  4. #include <linux/sched.h>
  5. #define kernel_locked() (current->lock_depth >= 0)
  6. extern int __lockfunc __reacquire_kernel_lock(void);
  7. extern void __lockfunc __release_kernel_lock(void);
  8. /*
  9. * Release/re-acquire global kernel lock for the scheduler
  10. */
  11. #define release_kernel_lock(tsk) do { \
  12. if (unlikely((tsk)->lock_depth >= 0)) \
  13. __release_kernel_lock(); \
  14. } while (0)
  15. static inline int reacquire_kernel_lock(struct task_struct *task)
  16. {
  17. if (unlikely(task->lock_depth >= 0))
  18. return __reacquire_kernel_lock();
  19. return 0;
  20. }
  21. extern void __lockfunc lock_kernel(void) __acquires(kernel_lock);
  22. extern void __lockfunc unlock_kernel(void) __releases(kernel_lock);
  23. /*
  24. * Various legacy drivers don't really need the BKL in a specific
  25. * function, but they *do* need to know that the BKL became available.
  26. * This function just avoids wrapping a bunch of lock/unlock pairs
  27. * around code which doesn't really need it.
  28. */
  29. static inline void cycle_kernel_lock(void)
  30. {
  31. lock_kernel();
  32. unlock_kernel();
  33. }
  34. #else
  35. #define lock_kernel() do { } while(0)
  36. #define unlock_kernel() do { } while(0)
  37. #define release_kernel_lock(task) do { } while(0)
  38. #define cycle_kernel_lock() do { } while(0)
  39. #define reacquire_kernel_lock(task) 0
  40. #define kernel_locked() 1
  41. #endif /* CONFIG_LOCK_KERNEL */
  42. #endif /* __LINUX_SMPLOCK_H */