|
@@ -467,7 +467,7 @@ void flush_tlb_all(void)
|
|
|
* it goes straight through and wastes no time serializing
|
|
|
* anything. Worst case is that we lose a reschedule ...
|
|
|
*/
|
|
|
-void native_smp_send_reschedule(int cpu)
|
|
|
+static void native_smp_send_reschedule(int cpu)
|
|
|
{
|
|
|
WARN_ON(cpu_is_offline(cpu));
|
|
|
send_IPI_mask(cpumask_of_cpu(cpu), RESCHEDULE_VECTOR);
|
|
@@ -546,9 +546,10 @@ static void __smp_call_function(void (*func) (void *info), void *info,
|
|
|
* You must not call this function with disabled interrupts or from a
|
|
|
* hardware interrupt handler or from a bottom half handler.
|
|
|
*/
|
|
|
-int native_smp_call_function_mask(cpumask_t mask,
|
|
|
- void (*func)(void *), void *info,
|
|
|
- int wait)
|
|
|
+static int
|
|
|
+native_smp_call_function_mask(cpumask_t mask,
|
|
|
+ void (*func)(void *), void *info,
|
|
|
+ int wait)
|
|
|
{
|
|
|
struct call_data_struct data;
|
|
|
cpumask_t allbutself;
|
|
@@ -599,60 +600,6 @@ int native_smp_call_function_mask(cpumask_t mask,
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-/**
|
|
|
- * smp_call_function(): Run a function on all other CPUs.
|
|
|
- * @func: The function to run. This must be fast and non-blocking.
|
|
|
- * @info: An arbitrary pointer to pass to the function.
|
|
|
- * @nonatomic: Unused.
|
|
|
- * @wait: If true, wait (atomically) until function has completed on other CPUs.
|
|
|
- *
|
|
|
- * Returns 0 on success, else a negative status code.
|
|
|
- *
|
|
|
- * If @wait is true, then returns once @func has returned; otherwise
|
|
|
- * it returns just before the target cpu calls @func.
|
|
|
- *
|
|
|
- * You must not call this function with disabled interrupts or from a
|
|
|
- * hardware interrupt handler or from a bottom half handler.
|
|
|
- */
|
|
|
-int smp_call_function(void (*func) (void *info), void *info, int nonatomic,
|
|
|
- int wait)
|
|
|
-{
|
|
|
- return smp_call_function_mask(cpu_online_map, func, info, wait);
|
|
|
-}
|
|
|
-EXPORT_SYMBOL(smp_call_function);
|
|
|
-
|
|
|
-/**
|
|
|
- * smp_call_function_single - Run a function on another CPU
|
|
|
- * @cpu: The target CPU. Cannot be the calling CPU.
|
|
|
- * @func: The function to run. This must be fast and non-blocking.
|
|
|
- * @info: An arbitrary pointer to pass to the function.
|
|
|
- * @nonatomic: Unused.
|
|
|
- * @wait: If true, wait until function has completed on other CPUs.
|
|
|
- *
|
|
|
- * Returns 0 on success, else a negative status code.
|
|
|
- *
|
|
|
- * If @wait is true, then returns once @func has returned; otherwise
|
|
|
- * it returns just before the target cpu calls @func.
|
|
|
- */
|
|
|
-int smp_call_function_single(int cpu, void (*func) (void *info), void *info,
|
|
|
- int nonatomic, int wait)
|
|
|
-{
|
|
|
- /* prevent preemption and reschedule on another processor */
|
|
|
- int ret;
|
|
|
- int me = get_cpu();
|
|
|
- if (cpu == me) {
|
|
|
- WARN_ON(1);
|
|
|
- put_cpu();
|
|
|
- return -EBUSY;
|
|
|
- }
|
|
|
-
|
|
|
- ret = smp_call_function_mask(cpumask_of_cpu(cpu), func, info, wait);
|
|
|
-
|
|
|
- put_cpu();
|
|
|
- return ret;
|
|
|
-}
|
|
|
-EXPORT_SYMBOL(smp_call_function_single);
|
|
|
-
|
|
|
static void stop_this_cpu (void * dummy)
|
|
|
{
|
|
|
local_irq_disable();
|
|
@@ -670,7 +617,7 @@ static void stop_this_cpu (void * dummy)
|
|
|
* this function calls the 'stop' function on all other CPUs in the system.
|
|
|
*/
|
|
|
|
|
|
-void native_smp_send_stop(void)
|
|
|
+static void native_smp_send_stop(void)
|
|
|
{
|
|
|
/* Don't deadlock on the call lock in panic */
|
|
|
int nolock = !spin_trylock(&call_lock);
|