|
@@ -46,6 +46,9 @@
|
|
|
|
|
|
#include "mce-internal.h"
|
|
|
|
|
|
+#define CREATE_TRACE_POINTS
|
|
|
+#include <trace/events/mce.h>
|
|
|
+
|
|
|
int mce_disabled __read_mostly;
|
|
|
|
|
|
#define MISC_MCELOG_MINOR 227
|
|
@@ -85,18 +88,26 @@ static DECLARE_WAIT_QUEUE_HEAD(mce_wait);
|
|
|
static DEFINE_PER_CPU(struct mce, mces_seen);
|
|
|
static int cpu_missing;
|
|
|
|
|
|
-static void default_decode_mce(struct mce *m)
|
|
|
+/*
|
|
|
+ * CPU/chipset specific EDAC code can register a notifier call here to print
|
|
|
+ * MCE errors in a human-readable form.
|
|
|
+ */
|
|
|
+ATOMIC_NOTIFIER_HEAD(x86_mce_decoder_chain);
|
|
|
+EXPORT_SYMBOL_GPL(x86_mce_decoder_chain);
|
|
|
+
|
|
|
+static int default_decode_mce(struct notifier_block *nb, unsigned long val,
|
|
|
+ void *data)
|
|
|
{
|
|
|
pr_emerg("No human readable MCE decoding support on this CPU type.\n");
|
|
|
pr_emerg("Run the message through 'mcelog --ascii' to decode.\n");
|
|
|
+
|
|
|
+ return NOTIFY_STOP;
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * CPU/chipset specific EDAC code can register a callback here to print
|
|
|
- * MCE errors in a human-readable form:
|
|
|
- */
|
|
|
-void (*x86_mce_decode_callback)(struct mce *m) = default_decode_mce;
|
|
|
-EXPORT_SYMBOL(x86_mce_decode_callback);
|
|
|
+static struct notifier_block mce_dec_nb = {
|
|
|
+ .notifier_call = default_decode_mce,
|
|
|
+ .priority = -1,
|
|
|
+};
|
|
|
|
|
|
/* MCA banks polled by the period polling timer for corrected events */
|
|
|
DEFINE_PER_CPU(mce_banks_t, mce_poll_banks) = {
|
|
@@ -141,6 +152,9 @@ void mce_log(struct mce *mce)
|
|
|
{
|
|
|
unsigned next, entry;
|
|
|
|
|
|
+ /* Emit the trace record: */
|
|
|
+ trace_mce_record(mce);
|
|
|
+
|
|
|
mce->finished = 0;
|
|
|
wmb();
|
|
|
for (;;) {
|
|
@@ -204,9 +218,9 @@ static void print_mce(struct mce *m)
|
|
|
|
|
|
/*
|
|
|
* Print out human-readable details about the MCE error,
|
|
|
- * (if the CPU has an implementation for that):
|
|
|
+ * (if the CPU has an implementation for that)
|
|
|
*/
|
|
|
- x86_mce_decode_callback(m);
|
|
|
+ atomic_notifier_call_chain(&x86_mce_decoder_chain, 0, m);
|
|
|
}
|
|
|
|
|
|
static void print_mce_head(void)
|
|
@@ -1122,7 +1136,7 @@ static int check_interval = 5 * 60; /* 5 minutes */
|
|
|
static DEFINE_PER_CPU(int, mce_next_interval); /* in jiffies */
|
|
|
static DEFINE_PER_CPU(struct timer_list, mce_timer);
|
|
|
|
|
|
-static void mcheck_timer(unsigned long data)
|
|
|
+static void mce_start_timer(unsigned long data)
|
|
|
{
|
|
|
struct timer_list *t = &per_cpu(mce_timer, data);
|
|
|
int *n;
|
|
@@ -1187,7 +1201,7 @@ int mce_notify_irq(void)
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(mce_notify_irq);
|
|
|
|
|
|
-static int mce_banks_init(void)
|
|
|
+static int __cpuinit __mcheck_cpu_mce_banks_init(void)
|
|
|
{
|
|
|
int i;
|
|
|
|
|
@@ -1206,7 +1220,7 @@ static int mce_banks_init(void)
|
|
|
/*
|
|
|
* Initialize Machine Checks for a CPU.
|
|
|
*/
|
|
|
-static int __cpuinit mce_cap_init(void)
|
|
|
+static int __cpuinit __mcheck_cpu_cap_init(void)
|
|
|
{
|
|
|
unsigned b;
|
|
|
u64 cap;
|
|
@@ -1228,7 +1242,7 @@ static int __cpuinit mce_cap_init(void)
|
|
|
WARN_ON(banks != 0 && b != banks);
|
|
|
banks = b;
|
|
|
if (!mce_banks) {
|
|
|
- int err = mce_banks_init();
|
|
|
+ int err = __mcheck_cpu_mce_banks_init();
|
|
|
|
|
|
if (err)
|
|
|
return err;
|
|
@@ -1244,7 +1258,7 @@ static int __cpuinit mce_cap_init(void)
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void mce_init(void)
|
|
|
+static void __mcheck_cpu_init_generic(void)
|
|
|
{
|
|
|
mce_banks_t all_banks;
|
|
|
u64 cap;
|
|
@@ -1273,7 +1287,7 @@ static void mce_init(void)
|
|
|
}
|
|
|
|
|
|
/* Add per CPU specific workarounds here */
|
|
|
-static int __cpuinit mce_cpu_quirks(struct cpuinfo_x86 *c)
|
|
|
+static int __cpuinit __mcheck_cpu_apply_quirks(struct cpuinfo_x86 *c)
|
|
|
{
|
|
|
if (c->x86_vendor == X86_VENDOR_UNKNOWN) {
|
|
|
pr_info("MCE: unknown CPU type - not enabling MCE support.\n");
|
|
@@ -1341,7 +1355,7 @@ static int __cpuinit mce_cpu_quirks(struct cpuinfo_x86 *c)
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void __cpuinit mce_ancient_init(struct cpuinfo_x86 *c)
|
|
|
+static void __cpuinit __mcheck_cpu_ancient_init(struct cpuinfo_x86 *c)
|
|
|
{
|
|
|
if (c->x86 != 5)
|
|
|
return;
|
|
@@ -1355,7 +1369,7 @@ static void __cpuinit mce_ancient_init(struct cpuinfo_x86 *c)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static void mce_cpu_features(struct cpuinfo_x86 *c)
|
|
|
+static void __mcheck_cpu_init_vendor(struct cpuinfo_x86 *c)
|
|
|
{
|
|
|
switch (c->x86_vendor) {
|
|
|
case X86_VENDOR_INTEL:
|
|
@@ -1369,7 +1383,7 @@ static void mce_cpu_features(struct cpuinfo_x86 *c)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static void mce_init_timer(void)
|
|
|
+static void __mcheck_cpu_init_timer(void)
|
|
|
{
|
|
|
struct timer_list *t = &__get_cpu_var(mce_timer);
|
|
|
int *n = &__get_cpu_var(mce_next_interval);
|
|
@@ -1380,7 +1394,7 @@ static void mce_init_timer(void)
|
|
|
*n = check_interval * HZ;
|
|
|
if (!*n)
|
|
|
return;
|
|
|
- setup_timer(t, mcheck_timer, smp_processor_id());
|
|
|
+ setup_timer(t, mce_start_timer, smp_processor_id());
|
|
|
t->expires = round_jiffies(jiffies + *n);
|
|
|
add_timer_on(t, smp_processor_id());
|
|
|
}
|
|
@@ -1400,27 +1414,28 @@ void (*machine_check_vector)(struct pt_regs *, long error_code) =
|
|
|
* Called for each booted CPU to set up machine checks.
|
|
|
* Must be called with preempt off:
|
|
|
*/
|
|
|
-void __cpuinit mcheck_init(struct cpuinfo_x86 *c)
|
|
|
+void __cpuinit mcheck_cpu_init(struct cpuinfo_x86 *c)
|
|
|
{
|
|
|
if (mce_disabled)
|
|
|
return;
|
|
|
|
|
|
- mce_ancient_init(c);
|
|
|
+ __mcheck_cpu_ancient_init(c);
|
|
|
|
|
|
if (!mce_available(c))
|
|
|
return;
|
|
|
|
|
|
- if (mce_cap_init() < 0 || mce_cpu_quirks(c) < 0) {
|
|
|
+ if (__mcheck_cpu_cap_init() < 0 || __mcheck_cpu_apply_quirks(c) < 0) {
|
|
|
mce_disabled = 1;
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
machine_check_vector = do_machine_check;
|
|
|
|
|
|
- mce_init();
|
|
|
- mce_cpu_features(c);
|
|
|
- mce_init_timer();
|
|
|
+ __mcheck_cpu_init_generic();
|
|
|
+ __mcheck_cpu_init_vendor(c);
|
|
|
+ __mcheck_cpu_init_timer();
|
|
|
INIT_WORK(&__get_cpu_var(mce_work), mce_process_work);
|
|
|
+
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -1640,6 +1655,15 @@ static int __init mcheck_enable(char *str)
|
|
|
}
|
|
|
__setup("mce", mcheck_enable);
|
|
|
|
|
|
+int __init mcheck_init(void)
|
|
|
+{
|
|
|
+ atomic_notifier_chain_register(&x86_mce_decoder_chain, &mce_dec_nb);
|
|
|
+
|
|
|
+ mcheck_intel_therm_init();
|
|
|
+
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* Sysfs support
|
|
|
*/
|
|
@@ -1648,7 +1672,7 @@ __setup("mce", mcheck_enable);
|
|
|
* Disable machine checks on suspend and shutdown. We can't really handle
|
|
|
* them later.
|
|
|
*/
|
|
|
-static int mce_disable(void)
|
|
|
+static int mce_disable_error_reporting(void)
|
|
|
{
|
|
|
int i;
|
|
|
|
|
@@ -1663,12 +1687,12 @@ static int mce_disable(void)
|
|
|
|
|
|
static int mce_suspend(struct sys_device *dev, pm_message_t state)
|
|
|
{
|
|
|
- return mce_disable();
|
|
|
+ return mce_disable_error_reporting();
|
|
|
}
|
|
|
|
|
|
static int mce_shutdown(struct sys_device *dev)
|
|
|
{
|
|
|
- return mce_disable();
|
|
|
+ return mce_disable_error_reporting();
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -1678,8 +1702,8 @@ static int mce_shutdown(struct sys_device *dev)
|
|
|
*/
|
|
|
static int mce_resume(struct sys_device *dev)
|
|
|
{
|
|
|
- mce_init();
|
|
|
- mce_cpu_features(¤t_cpu_data);
|
|
|
+ __mcheck_cpu_init_generic();
|
|
|
+ __mcheck_cpu_init_vendor(¤t_cpu_data);
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
@@ -1689,8 +1713,8 @@ static void mce_cpu_restart(void *data)
|
|
|
del_timer_sync(&__get_cpu_var(mce_timer));
|
|
|
if (!mce_available(¤t_cpu_data))
|
|
|
return;
|
|
|
- mce_init();
|
|
|
- mce_init_timer();
|
|
|
+ __mcheck_cpu_init_generic();
|
|
|
+ __mcheck_cpu_init_timer();
|
|
|
}
|
|
|
|
|
|
/* Reinit MCEs after user configuration changes */
|
|
@@ -1716,7 +1740,7 @@ static void mce_enable_ce(void *all)
|
|
|
cmci_reenable();
|
|
|
cmci_recheck();
|
|
|
if (all)
|
|
|
- mce_init_timer();
|
|
|
+ __mcheck_cpu_init_timer();
|
|
|
}
|
|
|
|
|
|
static struct sysdev_class mce_sysclass = {
|
|
@@ -1929,13 +1953,14 @@ static __cpuinit void mce_remove_device(unsigned int cpu)
|
|
|
}
|
|
|
|
|
|
/* Make sure there are no machine checks on offlined CPUs. */
|
|
|
-static void mce_disable_cpu(void *h)
|
|
|
+static void __cpuinit mce_disable_cpu(void *h)
|
|
|
{
|
|
|
unsigned long action = *(unsigned long *)h;
|
|
|
int i;
|
|
|
|
|
|
if (!mce_available(¤t_cpu_data))
|
|
|
return;
|
|
|
+
|
|
|
if (!(action & CPU_TASKS_FROZEN))
|
|
|
cmci_clear();
|
|
|
for (i = 0; i < banks; i++) {
|
|
@@ -1946,7 +1971,7 @@ static void mce_disable_cpu(void *h)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static void mce_reenable_cpu(void *h)
|
|
|
+static void __cpuinit mce_reenable_cpu(void *h)
|
|
|
{
|
|
|
unsigned long action = *(unsigned long *)h;
|
|
|
int i;
|
|
@@ -2025,7 +2050,7 @@ static __init void mce_init_banks(void)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static __init int mce_init_device(void)
|
|
|
+static __init int mcheck_init_device(void)
|
|
|
{
|
|
|
int err;
|
|
|
int i = 0;
|
|
@@ -2053,7 +2078,7 @@ static __init int mce_init_device(void)
|
|
|
return err;
|
|
|
}
|
|
|
|
|
|
-device_initcall(mce_init_device);
|
|
|
+device_initcall(mcheck_init_device);
|
|
|
|
|
|
/*
|
|
|
* Old style boot options parsing. Only for compatibility.
|
|
@@ -2101,7 +2126,7 @@ static int fake_panic_set(void *data, u64 val)
|
|
|
DEFINE_SIMPLE_ATTRIBUTE(fake_panic_fops, fake_panic_get,
|
|
|
fake_panic_set, "%llu\n");
|
|
|
|
|
|
-static int __init mce_debugfs_init(void)
|
|
|
+static int __init mcheck_debugfs_init(void)
|
|
|
{
|
|
|
struct dentry *dmce, *ffake_panic;
|
|
|
|
|
@@ -2115,5 +2140,5 @@ static int __init mce_debugfs_init(void)
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
|
-late_initcall(mce_debugfs_init);
|
|
|
+late_initcall(mcheck_debugfs_init);
|
|
|
#endif
|