tracehook.h 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576
  1. /*
  2. * Tracing hooks
  3. *
  4. * Copyright (C) 2008 Red Hat, Inc. All rights reserved.
  5. *
  6. * This copyrighted material is made available to anyone wishing to use,
  7. * modify, copy, or redistribute it subject to the terms and conditions
  8. * of the GNU General Public License v.2.
  9. *
  10. * This file defines hook entry points called by core code where
  11. * user tracing/debugging support might need to do something. These
  12. * entry points are called tracehook_*(). Each hook declared below
  13. * has a detailed kerneldoc comment giving the context (locking et
  14. * al) from which it is called, and the meaning of its return value.
  15. *
  16. * Each function here typically has only one call site, so it is ok
  17. * to have some nontrivial tracehook_*() inlines. In all cases, the
  18. * fast path when no tracing is enabled should be very short.
  19. *
  20. * The purpose of this file and the tracehook_* layer is to consolidate
  21. * the interface that the kernel core and arch code uses to enable any
  22. * user debugging or tracing facility (such as ptrace). The interfaces
  23. * here are carefully documented so that maintainers of core and arch
  24. * code do not need to think about the implementation details of the
  25. * tracing facilities. Likewise, maintainers of the tracing code do not
  26. * need to understand all the calling core or arch code in detail, just
  27. * documented circumstances of each call, such as locking conditions.
  28. *
  29. * If the calling core code changes so that locking is different, then
  30. * it is ok to change the interface documented here. The maintainer of
  31. * core code changing should notify the maintainers of the tracing code
  32. * that they need to work out the change.
  33. *
  34. * Some tracehook_*() inlines take arguments that the current tracing
  35. * implementations might not necessarily use. These function signatures
  36. * are chosen to pass in all the information that is on hand in the
  37. * caller and might conceivably be relevant to a tracer, so that the
  38. * core code won't have to be updated when tracing adds more features.
  39. * If a call site changes so that some of those parameters are no longer
  40. * already on hand without extra work, then the tracehook_* interface
  41. * can change so there is no make-work burden on the core code. The
  42. * maintainer of core code changing should notify the maintainers of the
  43. * tracing code that they need to work out the change.
  44. */
  45. #ifndef _LINUX_TRACEHOOK_H
  46. #define _LINUX_TRACEHOOK_H 1
  47. #include <linux/sched.h>
  48. #include <linux/ptrace.h>
  49. #include <linux/security.h>
  50. struct linux_binprm;
  51. /**
  52. * tracehook_expect_breakpoints - guess if task memory might be touched
  53. * @task: current task, making a new mapping
  54. *
  55. * Return nonzero if @task is expected to want breakpoint insertion in
  56. * its memory at some point. A zero return is no guarantee it won't
  57. * be done, but this is a hint that it's known to be likely.
  58. *
  59. * May be called with @task->mm->mmap_sem held for writing.
  60. */
  61. static inline int tracehook_expect_breakpoints(struct task_struct *task)
  62. {
  63. return (task_ptrace(task) & PT_PTRACED) != 0;
  64. }
  65. /*
  66. * ptrace report for syscall entry and exit looks identical.
  67. */
  68. static inline void ptrace_report_syscall(struct pt_regs *regs)
  69. {
  70. int ptrace = task_ptrace(current);
  71. if (!(ptrace & PT_PTRACED))
  72. return;
  73. ptrace_notify(SIGTRAP | ((ptrace & PT_TRACESYSGOOD) ? 0x80 : 0));
  74. /*
  75. * this isn't the same as continuing with a signal, but it will do
  76. * for normal use. strace only continues with a signal if the
  77. * stopping signal is not SIGTRAP. -brl
  78. */
  79. if (current->exit_code) {
  80. send_sig(current->exit_code, current, 1);
  81. current->exit_code = 0;
  82. }
  83. }
  84. /**
  85. * tracehook_report_syscall_entry - task is about to attempt a system call
  86. * @regs: user register state of current task
  87. *
  88. * This will be called if %TIF_SYSCALL_TRACE has been set, when the
  89. * current task has just entered the kernel for a system call.
  90. * Full user register state is available here. Changing the values
  91. * in @regs can affect the system call number and arguments to be tried.
  92. * It is safe to block here, preventing the system call from beginning.
  93. *
  94. * Returns zero normally, or nonzero if the calling arch code should abort
  95. * the system call. That must prevent normal entry so no system call is
  96. * made. If @task ever returns to user mode after this, its register state
  97. * is unspecified, but should be something harmless like an %ENOSYS error
  98. * return. It should preserve enough information so that syscall_rollback()
  99. * can work (see asm-generic/syscall.h).
  100. *
  101. * Called without locks, just after entering kernel mode.
  102. */
  103. static inline __must_check int tracehook_report_syscall_entry(
  104. struct pt_regs *regs)
  105. {
  106. ptrace_report_syscall(regs);
  107. return 0;
  108. }
  109. /**
  110. * tracehook_report_syscall_exit - task has just finished a system call
  111. * @regs: user register state of current task
  112. * @step: nonzero if simulating single-step or block-step
  113. *
  114. * This will be called if %TIF_SYSCALL_TRACE has been set, when the
  115. * current task has just finished an attempted system call. Full
  116. * user register state is available here. It is safe to block here,
  117. * preventing signals from being processed.
  118. *
  119. * If @step is nonzero, this report is also in lieu of the normal
  120. * trap that would follow the system call instruction because
  121. * user_enable_block_step() or user_enable_single_step() was used.
  122. * In this case, %TIF_SYSCALL_TRACE might not be set.
  123. *
  124. * Called without locks, just before checking for pending signals.
  125. */
  126. static inline void tracehook_report_syscall_exit(struct pt_regs *regs, int step)
  127. {
  128. ptrace_report_syscall(regs);
  129. }
  130. /**
  131. * tracehook_unsafe_exec - check for exec declared unsafe due to tracing
  132. * @task: current task doing exec
  133. *
  134. * Return %LSM_UNSAFE_* bits applied to an exec because of tracing.
  135. *
  136. * Called with task_lock() held on @task.
  137. */
  138. static inline int tracehook_unsafe_exec(struct task_struct *task)
  139. {
  140. int unsafe = 0;
  141. int ptrace = task_ptrace(task);
  142. if (ptrace & PT_PTRACED) {
  143. if (ptrace & PT_PTRACE_CAP)
  144. unsafe |= LSM_UNSAFE_PTRACE_CAP;
  145. else
  146. unsafe |= LSM_UNSAFE_PTRACE;
  147. }
  148. return unsafe;
  149. }
  150. /**
  151. * tracehook_tracer_task - return the task that is tracing the given task
  152. * @tsk: task to consider
  153. *
  154. * Returns NULL if noone is tracing @task, or the &struct task_struct
  155. * pointer to its tracer.
  156. *
  157. * Must called under rcu_read_lock(). The pointer returned might be kept
  158. * live only by RCU. During exec, this may be called with task_lock()
  159. * held on @task, still held from when tracehook_unsafe_exec() was called.
  160. */
  161. static inline struct task_struct *tracehook_tracer_task(struct task_struct *tsk)
  162. {
  163. if (task_ptrace(tsk) & PT_PTRACED)
  164. return rcu_dereference(tsk->parent);
  165. return NULL;
  166. }
  167. /**
  168. * tracehook_report_exec - a successful exec was completed
  169. * @fmt: &struct linux_binfmt that performed the exec
  170. * @bprm: &struct linux_binprm containing exec details
  171. * @regs: user-mode register state
  172. *
  173. * An exec just completed, we are shortly going to return to user mode.
  174. * The freshly initialized register state can be seen and changed in @regs.
  175. * The name, file and other pointers in @bprm are still on hand to be
  176. * inspected, but will be freed as soon as this returns.
  177. *
  178. * Called with no locks, but with some kernel resources held live
  179. * and a reference on @fmt->module.
  180. */
  181. static inline void tracehook_report_exec(struct linux_binfmt *fmt,
  182. struct linux_binprm *bprm,
  183. struct pt_regs *regs)
  184. {
  185. if (!ptrace_event(PT_TRACE_EXEC, PTRACE_EVENT_EXEC, 0) &&
  186. unlikely(task_ptrace(current) & PT_PTRACED))
  187. send_sig(SIGTRAP, current, 0);
  188. }
  189. /**
  190. * tracehook_report_exit - task has begun to exit
  191. * @exit_code: pointer to value destined for @current->exit_code
  192. *
  193. * @exit_code points to the value passed to do_exit(), which tracing
  194. * might change here. This is almost the first thing in do_exit(),
  195. * before freeing any resources or setting the %PF_EXITING flag.
  196. *
  197. * Called with no locks held.
  198. */
  199. static inline void tracehook_report_exit(long *exit_code)
  200. {
  201. ptrace_event(PT_TRACE_EXIT, PTRACE_EVENT_EXIT, *exit_code);
  202. }
  203. /**
  204. * tracehook_prepare_clone - prepare for new child to be cloned
  205. * @clone_flags: %CLONE_* flags from clone/fork/vfork system call
  206. *
  207. * This is called before a new user task is to be cloned.
  208. * Its return value will be passed to tracehook_finish_clone().
  209. *
  210. * Called with no locks held.
  211. */
  212. static inline int tracehook_prepare_clone(unsigned clone_flags)
  213. {
  214. if (clone_flags & CLONE_UNTRACED)
  215. return 0;
  216. if (clone_flags & CLONE_VFORK) {
  217. if (current->ptrace & PT_TRACE_VFORK)
  218. return PTRACE_EVENT_VFORK;
  219. } else if ((clone_flags & CSIGNAL) != SIGCHLD) {
  220. if (current->ptrace & PT_TRACE_CLONE)
  221. return PTRACE_EVENT_CLONE;
  222. } else if (current->ptrace & PT_TRACE_FORK)
  223. return PTRACE_EVENT_FORK;
  224. return 0;
  225. }
  226. /**
  227. * tracehook_finish_clone - new child created and being attached
  228. * @child: new child task
  229. * @clone_flags: %CLONE_* flags from clone/fork/vfork system call
  230. * @trace: return value from tracehook_prepare_clone()
  231. *
  232. * This is called immediately after adding @child to its parent's children list.
  233. * The @trace value is that returned by tracehook_prepare_clone().
  234. *
  235. * Called with current's siglock and write_lock_irq(&tasklist_lock) held.
  236. */
  237. static inline void tracehook_finish_clone(struct task_struct *child,
  238. unsigned long clone_flags, int trace)
  239. {
  240. ptrace_init_task(child, (clone_flags & CLONE_PTRACE) || trace);
  241. }
  242. /**
  243. * tracehook_report_clone - in parent, new child is about to start running
  244. * @regs: parent's user register state
  245. * @clone_flags: flags from parent's system call
  246. * @pid: new child's PID in the parent's namespace
  247. * @child: new child task
  248. *
  249. * Called after a child is set up, but before it has been started running.
  250. * This is not a good place to block, because the child has not started
  251. * yet. Suspend the child here if desired, and then block in
  252. * tracehook_report_clone_complete(). This must prevent the child from
  253. * self-reaping if tracehook_report_clone_complete() uses the @child
  254. * pointer; otherwise it might have died and been released by the time
  255. * tracehook_report_clone_complete() is called.
  256. *
  257. * Called with no locks held, but the child cannot run until this returns.
  258. */
  259. static inline void tracehook_report_clone(struct pt_regs *regs,
  260. unsigned long clone_flags,
  261. pid_t pid, struct task_struct *child)
  262. {
  263. if (unlikely(task_ptrace(child))) {
  264. /*
  265. * It doesn't matter who attached/attaching to this
  266. * task, the pending SIGSTOP is right in any case.
  267. */
  268. sigaddset(&child->pending.signal, SIGSTOP);
  269. set_tsk_thread_flag(child, TIF_SIGPENDING);
  270. }
  271. }
  272. /**
  273. * tracehook_report_clone_complete - new child is running
  274. * @trace: return value from tracehook_prepare_clone()
  275. * @regs: parent's user register state
  276. * @clone_flags: flags from parent's system call
  277. * @pid: new child's PID in the parent's namespace
  278. * @child: child task, already running
  279. *
  280. * This is called just after the child has started running. This is
  281. * just before the clone/fork syscall returns, or blocks for vfork
  282. * child completion if @clone_flags has the %CLONE_VFORK bit set.
  283. * The @child pointer may be invalid if a self-reaping child died and
  284. * tracehook_report_clone() took no action to prevent it from self-reaping.
  285. *
  286. * Called with no locks held.
  287. */
  288. static inline void tracehook_report_clone_complete(int trace,
  289. struct pt_regs *regs,
  290. unsigned long clone_flags,
  291. pid_t pid,
  292. struct task_struct *child)
  293. {
  294. if (unlikely(trace))
  295. ptrace_event(0, trace, pid);
  296. }
  297. /**
  298. * tracehook_report_vfork_done - vfork parent's child has exited or exec'd
  299. * @child: child task, already running
  300. * @pid: new child's PID in the parent's namespace
  301. *
  302. * Called after a %CLONE_VFORK parent has waited for the child to complete.
  303. * The clone/vfork system call will return immediately after this.
  304. * The @child pointer may be invalid if a self-reaping child died and
  305. * tracehook_report_clone() took no action to prevent it from self-reaping.
  306. *
  307. * Called with no locks held.
  308. */
  309. static inline void tracehook_report_vfork_done(struct task_struct *child,
  310. pid_t pid)
  311. {
  312. ptrace_event(PT_TRACE_VFORK_DONE, PTRACE_EVENT_VFORK_DONE, pid);
  313. }
  314. /**
  315. * tracehook_prepare_release_task - task is being reaped, clean up tracing
  316. * @task: task in %EXIT_DEAD state
  317. *
  318. * This is called in release_task() just before @task gets finally reaped
  319. * and freed. This would be the ideal place to remove and clean up any
  320. * tracing-related state for @task.
  321. *
  322. * Called with no locks held.
  323. */
  324. static inline void tracehook_prepare_release_task(struct task_struct *task)
  325. {
  326. }
  327. /**
  328. * tracehook_finish_release_task - final tracing clean-up
  329. * @task: task in %EXIT_DEAD state
  330. *
  331. * This is called in release_task() when @task is being in the middle of
  332. * being reaped. After this, there must be no tracing entanglements.
  333. *
  334. * Called with write_lock_irq(&tasklist_lock) held.
  335. */
  336. static inline void tracehook_finish_release_task(struct task_struct *task)
  337. {
  338. ptrace_release_task(task);
  339. }
  340. /**
  341. * tracehook_signal_handler - signal handler setup is complete
  342. * @sig: number of signal being delivered
  343. * @info: siginfo_t of signal being delivered
  344. * @ka: sigaction setting that chose the handler
  345. * @regs: user register state
  346. * @stepping: nonzero if debugger single-step or block-step in use
  347. *
  348. * Called by the arch code after a signal handler has been set up.
  349. * Register and stack state reflects the user handler about to run.
  350. * Signal mask changes have already been made.
  351. *
  352. * Called without locks, shortly before returning to user mode
  353. * (or handling more signals).
  354. */
  355. static inline void tracehook_signal_handler(int sig, siginfo_t *info,
  356. const struct k_sigaction *ka,
  357. struct pt_regs *regs, int stepping)
  358. {
  359. if (stepping)
  360. ptrace_notify(SIGTRAP);
  361. }
  362. /**
  363. * tracehook_consider_ignored_signal - suppress short-circuit of ignored signal
  364. * @task: task receiving the signal
  365. * @sig: signal number being sent
  366. *
  367. * Return zero iff tracing doesn't care to examine this ignored signal,
  368. * so it can short-circuit normal delivery and never even get queued.
  369. *
  370. * Called with @task->sighand->siglock held.
  371. */
  372. static inline int tracehook_consider_ignored_signal(struct task_struct *task,
  373. int sig)
  374. {
  375. return (task_ptrace(task) & PT_PTRACED) != 0;
  376. }
  377. /**
  378. * tracehook_consider_fatal_signal - suppress special handling of fatal signal
  379. * @task: task receiving the signal
  380. * @sig: signal number being sent
  381. *
  382. * Return nonzero to prevent special handling of this termination signal.
  383. * Normally handler for signal is %SIG_DFL. It can be %SIG_IGN if @sig is
  384. * ignored, in which case force_sig() is about to reset it to %SIG_DFL.
  385. * When this returns zero, this signal might cause a quick termination
  386. * that does not give the debugger a chance to intercept the signal.
  387. *
  388. * Called with or without @task->sighand->siglock held.
  389. */
  390. static inline int tracehook_consider_fatal_signal(struct task_struct *task,
  391. int sig)
  392. {
  393. return (task_ptrace(task) & PT_PTRACED) != 0;
  394. }
  395. /**
  396. * tracehook_force_sigpending - let tracing force signal_pending(current) on
  397. *
  398. * Called when recomputing our signal_pending() flag. Return nonzero
  399. * to force the signal_pending() flag on, so that tracehook_get_signal()
  400. * will be called before the next return to user mode.
  401. *
  402. * Called with @current->sighand->siglock held.
  403. */
  404. static inline int tracehook_force_sigpending(void)
  405. {
  406. return 0;
  407. }
  408. /**
  409. * tracehook_get_signal - deliver synthetic signal to traced task
  410. * @task: @current
  411. * @regs: task_pt_regs(@current)
  412. * @info: details of synthetic signal
  413. * @return_ka: sigaction for synthetic signal
  414. *
  415. * Return zero to check for a real pending signal normally.
  416. * Return -1 after releasing the siglock to repeat the check.
  417. * Return a signal number to induce an artifical signal delivery,
  418. * setting *@info and *@return_ka to specify its details and behavior.
  419. *
  420. * The @return_ka->sa_handler value controls the disposition of the
  421. * signal, no matter the signal number. For %SIG_DFL, the return value
  422. * is a representative signal to indicate the behavior (e.g. %SIGTERM
  423. * for death, %SIGQUIT for core dump, %SIGSTOP for job control stop,
  424. * %SIGTSTP for stop unless in an orphaned pgrp), but the signal number
  425. * reported will be @info->si_signo instead.
  426. *
  427. * Called with @task->sighand->siglock held, before dequeuing pending signals.
  428. */
  429. static inline int tracehook_get_signal(struct task_struct *task,
  430. struct pt_regs *regs,
  431. siginfo_t *info,
  432. struct k_sigaction *return_ka)
  433. {
  434. return 0;
  435. }
  436. /**
  437. * tracehook_notify_jctl - report about job control stop/continue
  438. * @notify: nonzero if this is the last thread in the group to stop
  439. * @why: %CLD_STOPPED or %CLD_CONTINUED
  440. *
  441. * This is called when we might call do_notify_parent_cldstop().
  442. * It's called when about to stop for job control; we are already in
  443. * %TASK_STOPPED state, about to call schedule(). It's also called when
  444. * a delayed %CLD_STOPPED or %CLD_CONTINUED report is ready to be made.
  445. *
  446. * Return nonzero to generate a %SIGCHLD with @why, which is
  447. * normal if @notify is nonzero.
  448. *
  449. * Called with no locks held.
  450. */
  451. static inline int tracehook_notify_jctl(int notify, int why)
  452. {
  453. return notify || (current->ptrace & PT_PTRACED);
  454. }
  455. #define DEATH_REAP -1
  456. #define DEATH_DELAYED_GROUP_LEADER -2
  457. /**
  458. * tracehook_notify_death - task is dead, ready to notify parent
  459. * @task: @current task now exiting
  460. * @death_cookie: value to pass to tracehook_report_death()
  461. * @group_dead: nonzero if this was the last thread in the group to die
  462. *
  463. * A return value >= 0 means call do_notify_parent() with that signal
  464. * number. Negative return value can be %DEATH_REAP to self-reap right
  465. * now, or %DEATH_DELAYED_GROUP_LEADER to a zombie without notifying our
  466. * parent. Note that a return value of 0 means a do_notify_parent() call
  467. * that sends no signal, but still wakes up a parent blocked in wait*().
  468. *
  469. * Called with write_lock_irq(&tasklist_lock) held.
  470. */
  471. static inline int tracehook_notify_death(struct task_struct *task,
  472. void **death_cookie, int group_dead)
  473. {
  474. if (task_detached(task))
  475. return task->ptrace ? SIGCHLD : DEATH_REAP;
  476. /*
  477. * If something other than our normal parent is ptracing us, then
  478. * send it a SIGCHLD instead of honoring exit_signal. exit_signal
  479. * only has special meaning to our real parent.
  480. */
  481. if (thread_group_empty(task) && !ptrace_reparented(task))
  482. return task->exit_signal;
  483. return task->ptrace ? SIGCHLD : DEATH_DELAYED_GROUP_LEADER;
  484. }
  485. /**
  486. * tracehook_report_death - task is dead and ready to be reaped
  487. * @task: @current task now exiting
  488. * @signal: return value from tracheook_notify_death()
  489. * @death_cookie: value passed back from tracehook_notify_death()
  490. * @group_dead: nonzero if this was the last thread in the group to die
  491. *
  492. * Thread has just become a zombie or is about to self-reap. If positive,
  493. * @signal is the signal number just sent to the parent (usually %SIGCHLD).
  494. * If @signal is %DEATH_REAP, this thread will self-reap. If @signal is
  495. * %DEATH_DELAYED_GROUP_LEADER, this is a delayed_group_leader() zombie.
  496. * The @death_cookie was passed back by tracehook_notify_death().
  497. *
  498. * If normal reaping is not inhibited, @task->exit_state might be changing
  499. * in parallel.
  500. *
  501. * Called without locks.
  502. */
  503. static inline void tracehook_report_death(struct task_struct *task,
  504. int signal, void *death_cookie,
  505. int group_dead)
  506. {
  507. }
  508. #ifdef TIF_NOTIFY_RESUME
  509. /**
  510. * set_notify_resume - cause tracehook_notify_resume() to be called
  511. * @task: task that will call tracehook_notify_resume()
  512. *
  513. * Calling this arranges that @task will call tracehook_notify_resume()
  514. * before returning to user mode. If it's already running in user mode,
  515. * it will enter the kernel and call tracehook_notify_resume() soon.
  516. * If it's blocked, it will not be woken.
  517. */
  518. static inline void set_notify_resume(struct task_struct *task)
  519. {
  520. if (!test_and_set_tsk_thread_flag(task, TIF_NOTIFY_RESUME))
  521. kick_process(task);
  522. }
  523. /**
  524. * tracehook_notify_resume - report when about to return to user mode
  525. * @regs: user-mode registers of @current task
  526. *
  527. * This is called when %TIF_NOTIFY_RESUME has been set. Now we are
  528. * about to return to user mode, and the user state in @regs can be
  529. * inspected or adjusted. The caller in arch code has cleared
  530. * %TIF_NOTIFY_RESUME before the call. If the flag gets set again
  531. * asynchronously, this will be called again before we return to
  532. * user mode.
  533. *
  534. * Called without locks.
  535. */
  536. static inline void tracehook_notify_resume(struct pt_regs *regs)
  537. {
  538. }
  539. #endif /* TIF_NOTIFY_RESUME */
  540. #endif /* <linux/tracehook.h> */