smp_lock.h 1.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354
  1. #ifndef __LINUX_SMPLOCK_H
  2. #define __LINUX_SMPLOCK_H
  3. #include <linux/config.h>
  4. #include <linux/sched.h>
  5. #include <linux/spinlock.h>
  6. #ifdef CONFIG_LOCK_KERNEL
  7. #define kernel_locked() (current->lock_depth >= 0)
  8. extern int __lockfunc __reacquire_kernel_lock(void);
  9. extern void __lockfunc __release_kernel_lock(void);
  10. /*
  11. * Release/re-acquire global kernel lock for the scheduler
  12. */
  13. #define release_kernel_lock(tsk) do { \
  14. if (unlikely((tsk)->lock_depth >= 0)) \
  15. __release_kernel_lock(); \
  16. } while (0)
  17. /*
  18. * Non-SMP kernels will never block on the kernel lock,
  19. * so we are better off returning a constant zero from
  20. * reacquire_kernel_lock() so that the compiler can see
  21. * it at compile-time.
  22. */
  23. #if defined(CONFIG_SMP) && !defined(CONFIG_PREEMPT_BKL)
  24. # define return_value_on_smp return
  25. #else
  26. # define return_value_on_smp
  27. #endif
  28. static inline int reacquire_kernel_lock(struct task_struct *task)
  29. {
  30. if (unlikely(task->lock_depth >= 0))
  31. return_value_on_smp __reacquire_kernel_lock();
  32. return 0;
  33. }
  34. extern void __lockfunc lock_kernel(void) __acquires(kernel_lock);
  35. extern void __lockfunc unlock_kernel(void) __releases(kernel_lock);
  36. #else
  37. #define lock_kernel() do { } while(0)
  38. #define unlock_kernel() do { } while(0)
  39. #define release_kernel_lock(task) do { } while(0)
  40. #define reacquire_kernel_lock(task) 0
  41. #define kernel_locked() 1
  42. #endif /* CONFIG_LOCK_KERNEL */
  43. #endif /* __LINUX_SMPLOCK_H */