|
@@ -24,8 +24,32 @@ extern int dump_fpu(struct pt_regs *, struct user_i387_struct *);
|
|
|
extern void math_state_restore(void);
|
|
|
|
|
|
extern bool irq_fpu_usable(void);
|
|
|
-extern void kernel_fpu_begin(void);
|
|
|
-extern void kernel_fpu_end(void);
|
|
|
+
|
|
|
+/*
|
|
|
+ * Careful: __kernel_fpu_begin/end() must be called with preempt disabled
|
|
|
+ * and they don't touch the preempt state on their own.
|
|
|
+ * If you enable preemption after __kernel_fpu_begin(), preempt notifier
|
|
|
+ * should call the __kernel_fpu_end() to prevent the kernel/user FPU
|
|
|
+ * state from getting corrupted. KVM for example uses this model.
|
|
|
+ *
|
|
|
+ * All other cases use kernel_fpu_begin/end() which disable preemption
|
|
|
+ * during kernel FPU usage.
|
|
|
+ */
|
|
|
+extern void __kernel_fpu_begin(void);
|
|
|
+extern void __kernel_fpu_end(void);
|
|
|
+
|
|
|
+static inline void kernel_fpu_begin(void)
|
|
|
+{
|
|
|
+ WARN_ON_ONCE(!irq_fpu_usable());
|
|
|
+ preempt_disable();
|
|
|
+ __kernel_fpu_begin();
|
|
|
+}
|
|
|
+
|
|
|
+static inline void kernel_fpu_end(void)
|
|
|
+{
|
|
|
+ __kernel_fpu_end();
|
|
|
+ preempt_enable();
|
|
|
+}
|
|
|
|
|
|
/*
|
|
|
* Some instructions like VIA's padlock instructions generate a spurious
|