|
@@ -1,15 +1,17 @@
|
|
|
/*
|
|
|
- * (c) 2005, 2006 Advanced Micro Devices, Inc.
|
|
|
+ * (c) 2005-2012 Advanced Micro Devices, Inc.
|
|
|
* Your use of this code is subject to the terms and conditions of the
|
|
|
* GNU general public license version 2. See "COPYING" or
|
|
|
* http://www.gnu.org/licenses/gpl.html
|
|
|
*
|
|
|
* Written by Jacob Shin - AMD, Inc.
|
|
|
*
|
|
|
- * Support : jacob.shin@amd.com
|
|
|
+ * Support: borislav.petkov@amd.com
|
|
|
*
|
|
|
* April 2006
|
|
|
* - added support for AMD Family 0x10 processors
|
|
|
+ * May 2012
|
|
|
+ * - major scrubbing
|
|
|
*
|
|
|
* All MC4_MISCi registers are shared between multi-cores
|
|
|
*/
|
|
@@ -25,6 +27,7 @@
|
|
|
#include <linux/cpu.h>
|
|
|
#include <linux/smp.h>
|
|
|
|
|
|
+#include <asm/amd_nb.h>
|
|
|
#include <asm/apic.h>
|
|
|
#include <asm/idle.h>
|
|
|
#include <asm/mce.h>
|
|
@@ -45,23 +48,15 @@
|
|
|
#define MASK_BLKPTR_LO 0xFF000000
|
|
|
#define MCG_XBLK_ADDR 0xC0000400
|
|
|
|
|
|
-struct threshold_block {
|
|
|
- unsigned int block;
|
|
|
- unsigned int bank;
|
|
|
- unsigned int cpu;
|
|
|
- u32 address;
|
|
|
- u16 interrupt_enable;
|
|
|
- bool interrupt_capable;
|
|
|
- u16 threshold_limit;
|
|
|
- struct kobject kobj;
|
|
|
- struct list_head miscj;
|
|
|
+static const char * const th_names[] = {
|
|
|
+ "load_store",
|
|
|
+ "insn_fetch",
|
|
|
+ "combined_unit",
|
|
|
+ "",
|
|
|
+ "northbridge",
|
|
|
+ "execution_unit",
|
|
|
};
|
|
|
|
|
|
-struct threshold_bank {
|
|
|
- struct kobject *kobj;
|
|
|
- struct threshold_block *blocks;
|
|
|
- cpumask_var_t cpus;
|
|
|
-};
|
|
|
static DEFINE_PER_CPU(struct threshold_bank * [NR_BANKS], threshold_banks);
|
|
|
|
|
|
static unsigned char shared_bank[NR_BANKS] = {
|
|
@@ -84,6 +79,26 @@ struct thresh_restart {
|
|
|
u16 old_limit;
|
|
|
};
|
|
|
|
|
|
+static const char * const bank4_names(struct threshold_block *b)
|
|
|
+{
|
|
|
+ switch (b->address) {
|
|
|
+ /* MSR4_MISC0 */
|
|
|
+ case 0x00000413:
|
|
|
+ return "dram";
|
|
|
+
|
|
|
+ case 0xc0000408:
|
|
|
+ return "ht_links";
|
|
|
+
|
|
|
+ case 0xc0000409:
|
|
|
+ return "l3_cache";
|
|
|
+
|
|
|
+ default:
|
|
|
+ WARN(1, "Funny MSR: 0x%08x\n", b->address);
|
|
|
+ return "";
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+
|
|
|
static bool lvt_interrupt_supported(unsigned int bank, u32 msr_high_bits)
|
|
|
{
|
|
|
/*
|
|
@@ -224,8 +239,6 @@ void mce_amd_feature_init(struct cpuinfo_x86 *c)
|
|
|
|
|
|
if (!block)
|
|
|
per_cpu(bank_map, cpu) |= (1 << bank);
|
|
|
- if (shared_bank[bank] && c->cpu_core_id)
|
|
|
- break;
|
|
|
|
|
|
memset(&b, 0, sizeof(b));
|
|
|
b.cpu = cpu;
|
|
@@ -326,7 +339,7 @@ struct threshold_attr {
|
|
|
#define SHOW_FIELDS(name) \
|
|
|
static ssize_t show_ ## name(struct threshold_block *b, char *buf) \
|
|
|
{ \
|
|
|
- return sprintf(buf, "%lx\n", (unsigned long) b->name); \
|
|
|
+ return sprintf(buf, "%lu\n", (unsigned long) b->name); \
|
|
|
}
|
|
|
SHOW_FIELDS(interrupt_enable)
|
|
|
SHOW_FIELDS(threshold_limit)
|
|
@@ -377,38 +390,21 @@ store_threshold_limit(struct threshold_block *b, const char *buf, size_t size)
|
|
|
return size;
|
|
|
}
|
|
|
|
|
|
-struct threshold_block_cross_cpu {
|
|
|
- struct threshold_block *tb;
|
|
|
- long retval;
|
|
|
-};
|
|
|
-
|
|
|
-static void local_error_count_handler(void *_tbcc)
|
|
|
-{
|
|
|
- struct threshold_block_cross_cpu *tbcc = _tbcc;
|
|
|
- struct threshold_block *b = tbcc->tb;
|
|
|
- u32 low, high;
|
|
|
-
|
|
|
- rdmsr(b->address, low, high);
|
|
|
- tbcc->retval = (high & 0xFFF) - (THRESHOLD_MAX - b->threshold_limit);
|
|
|
-}
|
|
|
-
|
|
|
static ssize_t show_error_count(struct threshold_block *b, char *buf)
|
|
|
{
|
|
|
- struct threshold_block_cross_cpu tbcc = { .tb = b, };
|
|
|
+ u32 lo, hi;
|
|
|
|
|
|
- smp_call_function_single(b->cpu, local_error_count_handler, &tbcc, 1);
|
|
|
- return sprintf(buf, "%lx\n", tbcc.retval);
|
|
|
-}
|
|
|
+ rdmsr_on_cpu(b->cpu, b->address, &lo, &hi);
|
|
|
|
|
|
-static ssize_t store_error_count(struct threshold_block *b,
|
|
|
- const char *buf, size_t count)
|
|
|
-{
|
|
|
- struct thresh_restart tr = { .b = b, .reset = 1, .old_limit = 0 };
|
|
|
-
|
|
|
- smp_call_function_single(b->cpu, threshold_restart_bank, &tr, 1);
|
|
|
- return 1;
|
|
|
+ return sprintf(buf, "%u\n", ((hi & THRESHOLD_MAX) -
|
|
|
+ (THRESHOLD_MAX - b->threshold_limit)));
|
|
|
}
|
|
|
|
|
|
+static struct threshold_attr error_count = {
|
|
|
+ .attr = {.name = __stringify(error_count), .mode = 0444 },
|
|
|
+ .show = show_error_count,
|
|
|
+};
|
|
|
+
|
|
|
#define RW_ATTR(val) \
|
|
|
static struct threshold_attr val = { \
|
|
|
.attr = {.name = __stringify(val), .mode = 0644 }, \
|
|
@@ -418,7 +414,6 @@ static struct threshold_attr val = { \
|
|
|
|
|
|
RW_ATTR(interrupt_enable);
|
|
|
RW_ATTR(threshold_limit);
|
|
|
-RW_ATTR(error_count);
|
|
|
|
|
|
static struct attribute *default_attrs[] = {
|
|
|
&threshold_limit.attr,
|
|
@@ -517,7 +512,7 @@ static __cpuinit int allocate_threshold_blocks(unsigned int cpu,
|
|
|
|
|
|
err = kobject_init_and_add(&b->kobj, &threshold_ktype,
|
|
|
per_cpu(threshold_banks, cpu)[bank]->kobj,
|
|
|
- "misc%i", block);
|
|
|
+ (bank == 4 ? bank4_names(b) : th_names[bank]));
|
|
|
if (err)
|
|
|
goto out_free;
|
|
|
recurse:
|
|
@@ -548,98 +543,91 @@ out_free:
|
|
|
return err;
|
|
|
}
|
|
|
|
|
|
-static __cpuinit long
|
|
|
-local_allocate_threshold_blocks(int cpu, unsigned int bank)
|
|
|
+static __cpuinit int __threshold_add_blocks(struct threshold_bank *b)
|
|
|
{
|
|
|
- return allocate_threshold_blocks(cpu, bank, 0,
|
|
|
- MSR_IA32_MC0_MISC + bank * 4);
|
|
|
+ struct list_head *head = &b->blocks->miscj;
|
|
|
+ struct threshold_block *pos = NULL;
|
|
|
+ struct threshold_block *tmp = NULL;
|
|
|
+ int err = 0;
|
|
|
+
|
|
|
+ err = kobject_add(&b->blocks->kobj, b->kobj, b->blocks->kobj.name);
|
|
|
+ if (err)
|
|
|
+ return err;
|
|
|
+
|
|
|
+ list_for_each_entry_safe(pos, tmp, head, miscj) {
|
|
|
+
|
|
|
+ err = kobject_add(&pos->kobj, b->kobj, pos->kobj.name);
|
|
|
+ if (err) {
|
|
|
+ list_for_each_entry_safe_reverse(pos, tmp, head, miscj)
|
|
|
+ kobject_del(&pos->kobj);
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return err;
|
|
|
}
|
|
|
|
|
|
-/* symlinks sibling shared banks to first core. first core owns dir/files. */
|
|
|
static __cpuinit int threshold_create_bank(unsigned int cpu, unsigned int bank)
|
|
|
{
|
|
|
- int i, err = 0;
|
|
|
- struct threshold_bank *b = NULL;
|
|
|
struct device *dev = per_cpu(mce_device, cpu);
|
|
|
- char name[32];
|
|
|
-
|
|
|
- sprintf(name, "threshold_bank%i", bank);
|
|
|
+ struct amd_northbridge *nb = NULL;
|
|
|
+ struct threshold_bank *b = NULL;
|
|
|
+ const char *name = th_names[bank];
|
|
|
+ int err = 0;
|
|
|
|
|
|
-#ifdef CONFIG_SMP
|
|
|
- if (cpu_data(cpu).cpu_core_id && shared_bank[bank]) { /* symlink */
|
|
|
- i = cpumask_first(cpu_llc_shared_mask(cpu));
|
|
|
+ if (shared_bank[bank]) {
|
|
|
|
|
|
- /* first core not up yet */
|
|
|
- if (cpu_data(i).cpu_core_id)
|
|
|
- goto out;
|
|
|
+ nb = node_to_amd_nb(amd_get_nb_id(cpu));
|
|
|
+ WARN_ON(!nb);
|
|
|
|
|
|
- /* already linked */
|
|
|
- if (per_cpu(threshold_banks, cpu)[bank])
|
|
|
- goto out;
|
|
|
+ /* threshold descriptor already initialized on this node? */
|
|
|
+ if (nb->bank4) {
|
|
|
+ /* yes, use it */
|
|
|
+ b = nb->bank4;
|
|
|
+ err = kobject_add(b->kobj, &dev->kobj, name);
|
|
|
+ if (err)
|
|
|
+ goto out;
|
|
|
|
|
|
- b = per_cpu(threshold_banks, i)[bank];
|
|
|
+ per_cpu(threshold_banks, cpu)[bank] = b;
|
|
|
+ atomic_inc(&b->cpus);
|
|
|
|
|
|
- if (!b)
|
|
|
- goto out;
|
|
|
+ err = __threshold_add_blocks(b);
|
|
|
|
|
|
- err = sysfs_create_link(&dev->kobj, b->kobj, name);
|
|
|
- if (err)
|
|
|
goto out;
|
|
|
-
|
|
|
- cpumask_copy(b->cpus, cpu_llc_shared_mask(cpu));
|
|
|
- per_cpu(threshold_banks, cpu)[bank] = b;
|
|
|
-
|
|
|
- goto out;
|
|
|
+ }
|
|
|
}
|
|
|
-#endif
|
|
|
|
|
|
b = kzalloc(sizeof(struct threshold_bank), GFP_KERNEL);
|
|
|
if (!b) {
|
|
|
err = -ENOMEM;
|
|
|
goto out;
|
|
|
}
|
|
|
- if (!zalloc_cpumask_var(&b->cpus, GFP_KERNEL)) {
|
|
|
- kfree(b);
|
|
|
- err = -ENOMEM;
|
|
|
- goto out;
|
|
|
- }
|
|
|
|
|
|
b->kobj = kobject_create_and_add(name, &dev->kobj);
|
|
|
- if (!b->kobj)
|
|
|
+ if (!b->kobj) {
|
|
|
+ err = -EINVAL;
|
|
|
goto out_free;
|
|
|
-
|
|
|
-#ifndef CONFIG_SMP
|
|
|
- cpumask_setall(b->cpus);
|
|
|
-#else
|
|
|
- cpumask_set_cpu(cpu, b->cpus);
|
|
|
-#endif
|
|
|
+ }
|
|
|
|
|
|
per_cpu(threshold_banks, cpu)[bank] = b;
|
|
|
|
|
|
- err = local_allocate_threshold_blocks(cpu, bank);
|
|
|
- if (err)
|
|
|
- goto out_free;
|
|
|
-
|
|
|
- for_each_cpu(i, b->cpus) {
|
|
|
- if (i == cpu)
|
|
|
- continue;
|
|
|
+ if (shared_bank[bank]) {
|
|
|
+ atomic_set(&b->cpus, 1);
|
|
|
|
|
|
- dev = per_cpu(mce_device, i);
|
|
|
- if (dev)
|
|
|
- err = sysfs_create_link(&dev->kobj,b->kobj, name);
|
|
|
- if (err)
|
|
|
- goto out;
|
|
|
-
|
|
|
- per_cpu(threshold_banks, i)[bank] = b;
|
|
|
+ /* nb is already initialized, see above */
|
|
|
+ WARN_ON(nb->bank4);
|
|
|
+ nb->bank4 = b;
|
|
|
}
|
|
|
|
|
|
- goto out;
|
|
|
+ err = allocate_threshold_blocks(cpu, bank, 0,
|
|
|
+ MSR_IA32_MC0_MISC + bank * 4);
|
|
|
+ if (!err)
|
|
|
+ goto out;
|
|
|
|
|
|
-out_free:
|
|
|
- per_cpu(threshold_banks, cpu)[bank] = NULL;
|
|
|
- free_cpumask_var(b->cpus);
|
|
|
+ out_free:
|
|
|
kfree(b);
|
|
|
-out:
|
|
|
+
|
|
|
+ out:
|
|
|
return err;
|
|
|
}
|
|
|
|
|
@@ -660,12 +648,6 @@ static __cpuinit int threshold_create_device(unsigned int cpu)
|
|
|
return err;
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * let's be hotplug friendly.
|
|
|
- * in case of multiple core processors, the first core always takes ownership
|
|
|
- * of shared sysfs dir/files, and rest of the cores will be symlinked to it.
|
|
|
- */
|
|
|
-
|
|
|
static void deallocate_threshold_block(unsigned int cpu,
|
|
|
unsigned int bank)
|
|
|
{
|
|
@@ -686,41 +668,42 @@ static void deallocate_threshold_block(unsigned int cpu,
|
|
|
per_cpu(threshold_banks, cpu)[bank]->blocks = NULL;
|
|
|
}
|
|
|
|
|
|
+static void __threshold_remove_blocks(struct threshold_bank *b)
|
|
|
+{
|
|
|
+ struct threshold_block *pos = NULL;
|
|
|
+ struct threshold_block *tmp = NULL;
|
|
|
+
|
|
|
+ kobject_del(b->kobj);
|
|
|
+
|
|
|
+ list_for_each_entry_safe(pos, tmp, &b->blocks->miscj, miscj)
|
|
|
+ kobject_del(&pos->kobj);
|
|
|
+}
|
|
|
+
|
|
|
static void threshold_remove_bank(unsigned int cpu, int bank)
|
|
|
{
|
|
|
+ struct amd_northbridge *nb;
|
|
|
struct threshold_bank *b;
|
|
|
- struct device *dev;
|
|
|
- char name[32];
|
|
|
- int i = 0;
|
|
|
|
|
|
b = per_cpu(threshold_banks, cpu)[bank];
|
|
|
if (!b)
|
|
|
return;
|
|
|
+
|
|
|
if (!b->blocks)
|
|
|
goto free_out;
|
|
|
|
|
|
- sprintf(name, "threshold_bank%i", bank);
|
|
|
-
|
|
|
-#ifdef CONFIG_SMP
|
|
|
- /* sibling symlink */
|
|
|
- if (shared_bank[bank] && b->blocks->cpu != cpu) {
|
|
|
- dev = per_cpu(mce_device, cpu);
|
|
|
- sysfs_remove_link(&dev->kobj, name);
|
|
|
- per_cpu(threshold_banks, cpu)[bank] = NULL;
|
|
|
-
|
|
|
- return;
|
|
|
- }
|
|
|
-#endif
|
|
|
-
|
|
|
- /* remove all sibling symlinks before unregistering */
|
|
|
- for_each_cpu(i, b->cpus) {
|
|
|
- if (i == cpu)
|
|
|
- continue;
|
|
|
-
|
|
|
- dev = per_cpu(mce_device, i);
|
|
|
- if (dev)
|
|
|
- sysfs_remove_link(&dev->kobj, name);
|
|
|
- per_cpu(threshold_banks, i)[bank] = NULL;
|
|
|
+ if (shared_bank[bank]) {
|
|
|
+ if (!atomic_dec_and_test(&b->cpus)) {
|
|
|
+ __threshold_remove_blocks(b);
|
|
|
+ per_cpu(threshold_banks, cpu)[bank] = NULL;
|
|
|
+ return;
|
|
|
+ } else {
|
|
|
+ /*
|
|
|
+ * the last CPU on this node using the shared bank is
|
|
|
+ * going away, remove that bank now.
|
|
|
+ */
|
|
|
+ nb = node_to_amd_nb(amd_get_nb_id(cpu));
|
|
|
+ nb->bank4 = NULL;
|
|
|
+ }
|
|
|
}
|
|
|
|
|
|
deallocate_threshold_block(cpu, bank);
|
|
@@ -728,7 +711,6 @@ static void threshold_remove_bank(unsigned int cpu, int bank)
|
|
|
free_out:
|
|
|
kobject_del(b->kobj);
|
|
|
kobject_put(b->kobj);
|
|
|
- free_cpumask_var(b->cpus);
|
|
|
kfree(b);
|
|
|
per_cpu(threshold_banks, cpu)[bank] = NULL;
|
|
|
}
|