process.c 4.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187
  1. /*
  2. * Copyright (C) 2000-2003 Axis Communications AB
  3. *
  4. * Authors: Bjorn Wesen (bjornw@axis.com)
  5. * Mikael Starvik (starvik@axis.com)
  6. * Tobias Anderberg (tobiasa@axis.com), CRISv32 port.
  7. *
  8. * This file handles the architecture-dependent parts of process handling..
  9. */
  10. #include <linux/sched.h>
  11. #include <linux/slab.h>
  12. #include <linux/err.h>
  13. #include <linux/fs.h>
  14. #include <hwregs/reg_rdwr.h>
  15. #include <hwregs/reg_map.h>
  16. #include <hwregs/timer_defs.h>
  17. #include <hwregs/intr_vect_defs.h>
  18. #include <linux/ptrace.h>
  19. extern void stop_watchdog(void);
  20. extern int cris_hlt_counter;
  21. /* We use this if we don't have any better idle routine. */
  22. void default_idle(void)
  23. {
  24. local_irq_disable();
  25. if (!need_resched() && !cris_hlt_counter) {
  26. /* Halt until exception. */
  27. __asm__ volatile("ei \n\t"
  28. "halt ");
  29. }
  30. local_irq_enable();
  31. }
  32. /*
  33. * Free current thread data structures etc..
  34. */
  35. extern void deconfigure_bp(long pid);
  36. void exit_thread(void)
  37. {
  38. deconfigure_bp(current->pid);
  39. }
  40. /*
  41. * If the watchdog is enabled, disable interrupts and enter an infinite loop.
  42. * The watchdog will reset the CPU after 0.1s. If the watchdog isn't enabled
  43. * then enable it and wait.
  44. */
  45. extern void arch_enable_nmi(void);
  46. void
  47. hard_reset_now(void)
  48. {
  49. /*
  50. * Don't declare this variable elsewhere. We don't want any other
  51. * code to know about it than the watchdog handler in entry.S and
  52. * this code, implementing hard reset through the watchdog.
  53. */
  54. #if defined(CONFIG_ETRAX_WATCHDOG)
  55. extern int cause_of_death;
  56. #endif
  57. printk("*** HARD RESET ***\n");
  58. local_irq_disable();
  59. #if defined(CONFIG_ETRAX_WATCHDOG)
  60. cause_of_death = 0xbedead;
  61. #else
  62. {
  63. reg_timer_rw_wd_ctrl wd_ctrl = {0};
  64. stop_watchdog();
  65. wd_ctrl.key = 16; /* Arbitrary key. */
  66. wd_ctrl.cnt = 1; /* Minimum time. */
  67. wd_ctrl.cmd = regk_timer_start;
  68. arch_enable_nmi();
  69. REG_WR(timer, regi_timer0, rw_wd_ctrl, wd_ctrl);
  70. }
  71. #endif
  72. while (1)
  73. ; /* Wait for reset. */
  74. }
  75. /*
  76. * Return saved PC of a blocked thread.
  77. */
  78. unsigned long thread_saved_pc(struct task_struct *t)
  79. {
  80. return task_pt_regs(t)->erp;
  81. }
  82. /*
  83. * Setup the child's kernel stack with a pt_regs and call switch_stack() on it.
  84. * It will be unnested during _resume and _ret_from_sys_call when the new thread
  85. * is scheduled.
  86. *
  87. * Also setup the thread switching structure which is used to keep
  88. * thread-specific data during _resumes.
  89. */
  90. extern asmlinkage void ret_from_fork(void);
  91. extern asmlinkage void ret_from_kernel_thread(void);
  92. int
  93. copy_thread(unsigned long clone_flags, unsigned long usp,
  94. unsigned long arg, struct task_struct *p)
  95. {
  96. struct pt_regs *childregs = task_pt_regs(p);
  97. struct switch_stack *swstack = ((struct switch_stack *) childregs) - 1;
  98. /*
  99. * Put the pt_regs structure at the end of the new kernel stack page and
  100. * fix it up. Note: the task_struct doubles as the kernel stack for the
  101. * task.
  102. */
  103. if (unlikely(p->flags & PF_KTHREAD)) {
  104. memset(swstack, 0,
  105. sizeof(struct switch_stack) + sizeof(struct pt_regs));
  106. swstack->r1 = usp;
  107. swstack->r2 = arg;
  108. childregs->ccs = 1 << (I_CCS_BITNR + CCS_SHIFT);
  109. swstack->return_ip = (unsigned long) ret_from_kernel_thread;
  110. p->thread.ksp = (unsigned long) swstack;
  111. p->thread.usp = 0;
  112. return 0;
  113. }
  114. *childregs = *current_pt_regs(); /* Struct copy of pt_regs. */
  115. childregs->r10 = 0; /* Child returns 0 after a fork/clone. */
  116. /* Set a new TLS ?
  117. * The TLS is in $mof because it is the 5th argument to sys_clone.
  118. */
  119. if (p->mm && (clone_flags & CLONE_SETTLS)) {
  120. task_thread_info(p)->tls = childregs->mof;
  121. }
  122. /* Put the switch stack right below the pt_regs. */
  123. /* Parameter to ret_from_sys_call. 0 is don't restart the syscall. */
  124. swstack->r9 = 0;
  125. /*
  126. * We want to return into ret_from_sys_call after the _resume.
  127. * ret_from_fork will call ret_from_sys_call.
  128. */
  129. swstack->return_ip = (unsigned long) ret_from_fork;
  130. /* Fix the user-mode and kernel-mode stackpointer. */
  131. p->thread.usp = usp ?: rdusp();
  132. p->thread.ksp = (unsigned long) swstack;
  133. return 0;
  134. }
  135. unsigned long
  136. get_wchan(struct task_struct *p)
  137. {
  138. /* TODO */
  139. return 0;
  140. }
  141. #undef last_sched
  142. #undef first_sched
  143. void show_regs(struct pt_regs * regs)
  144. {
  145. unsigned long usp = rdusp();
  146. printk("ERP: %08lx SRP: %08lx CCS: %08lx USP: %08lx MOF: %08lx\n",
  147. regs->erp, regs->srp, regs->ccs, usp, regs->mof);
  148. printk(" r0: %08lx r1: %08lx r2: %08lx r3: %08lx\n",
  149. regs->r0, regs->r1, regs->r2, regs->r3);
  150. printk(" r4: %08lx r5: %08lx r6: %08lx r7: %08lx\n",
  151. regs->r4, regs->r5, regs->r6, regs->r7);
  152. printk(" r8: %08lx r9: %08lx r10: %08lx r11: %08lx\n",
  153. regs->r8, regs->r9, regs->r10, regs->r11);
  154. printk("r12: %08lx r13: %08lx oR10: %08lx\n",
  155. regs->r12, regs->r13, regs->orig_r10);
  156. }