Merge branch 'linus' into x86/urgent
authorIngo Molnar <mingo@elte.hu>
Thu, 24 Sep 2009 10:59:11 +0000 (12:59 +0200)
committerIngo Molnar <mingo@elte.hu>
Thu, 24 Sep 2009 10:59:18 +0000 (12:59 +0200)
Merge reason: Queueing up dependent early-printk fix.

Signed-off-by: Ingo Molnar <mingo@elte.hu>
arch/x86/include/asm/pgtable_types.h
arch/x86/kernel/cpu/mcheck/mce-inject.c
arch/x86/kernel/cpu/mcheck/mce.c
arch/x86/kernel/cpu/mcheck/therm_throt.c
arch/x86/kernel/tsc_sync.c
arch/x86/mm/Makefile
arch/x86/mm/init.c
arch/x86/mm/pat.c
arch/x86/mm/setup_nx.c [new file with mode: 0644]
arch/x86/xen/enlighten.c

index 7b467bf3c68067218d1a79a40818018b9ed90134..d1f4a760be23dc6cfe9a2feff61a7b07d1dc8c46 100644 (file)
@@ -277,6 +277,7 @@ static inline pteval_t pte_flags(pte_t pte)
 typedef struct page *pgtable_t;
 
 extern pteval_t __supported_pte_mask;
+extern void set_nx(void);
 extern int nx_enabled;
 
 #define pgprot_writecombine    pgprot_writecombine
index 7029f0e2acadef9bc0e6d93820cc4ab08bb09797..472763d9209841dd247c9052bb29144b8682b097 100644 (file)
@@ -98,8 +98,9 @@ static struct notifier_block mce_raise_nb = {
 };
 
 /* Inject mce on current CPU */
-static int raise_local(struct mce *m)
+static int raise_local(void)
 {
+       struct mce *m = &__get_cpu_var(injectm);
        int context = MCJ_CTX(m->inject_flags);
        int ret = 0;
        int cpu = m->extcpu;
@@ -167,12 +168,12 @@ static void raise_mce(struct mce *m)
                        }
                        cpu_relax();
                }
-               raise_local(m);
+               raise_local();
                put_cpu();
                put_online_cpus();
        } else
 #endif
-               raise_local(m);
+               raise_local();
 }
 
 /* Error injection interface */
index 2f5aab26320e120b3531247e7682b1221721f96e..4b2af86e3e8d2388a74b2dec2b1145253b43427a 100644 (file)
@@ -305,13 +305,25 @@ static int msr_to_offset(u32 msr)
 static u64 mce_rdmsrl(u32 msr)
 {
        u64 v;
+
        if (__get_cpu_var(injectm).finished) {
                int offset = msr_to_offset(msr);
+
                if (offset < 0)
                        return 0;
                return *(u64 *)((char *)&__get_cpu_var(injectm) + offset);
        }
-       rdmsrl(msr, v);
+
+       if (rdmsrl_safe(msr, &v)) {
+               WARN_ONCE(1, "mce: Unable to read msr %d!\n", msr);
+               /*
+                * Return zero in case the access faulted. This should
+                * not happen normally but can happen if the CPU does
+                * something weird, or if the code is buggy.
+                */
+               v = 0;
+       }
+
        return v;
 }
 
@@ -319,6 +331,7 @@ static void mce_wrmsrl(u32 msr, u64 v)
 {
        if (__get_cpu_var(injectm).finished) {
                int offset = msr_to_offset(msr);
+
                if (offset >= 0)
                        *(u64 *)((char *)&__get_cpu_var(injectm) + offset) = v;
                return;
@@ -415,7 +428,7 @@ static inline void mce_get_rip(struct mce *m, struct pt_regs *regs)
                m->ip = mce_rdmsrl(rip_msr);
 }
 
-#ifdef CONFIG_X86_LOCAL_APIC 
+#ifdef CONFIG_X86_LOCAL_APIC
 /*
  * Called after interrupts have been reenabled again
  * when a MCE happened during an interrupts off region
@@ -1172,6 +1185,7 @@ static int mce_banks_init(void)
                return -ENOMEM;
        for (i = 0; i < banks; i++) {
                struct mce_bank *b = &mce_banks[i];
+
                b->ctl = -1ULL;
                b->init = 1;
        }
@@ -1203,6 +1217,7 @@ static int __cpuinit mce_cap_init(void)
        banks = b;
        if (!mce_banks) {
                int err = mce_banks_init();
+
                if (err)
                        return err;
        }
@@ -1237,6 +1252,7 @@ static void mce_init(void)
 
        for (i = 0; i < banks; i++) {
                struct mce_bank *b = &mce_banks[i];
+
                if (!b->init)
                        continue;
                wrmsrl(MSR_IA32_MCx_CTL(i), b->ctl);
@@ -1626,6 +1642,7 @@ static int mce_disable(void)
 
        for (i = 0; i < banks; i++) {
                struct mce_bank *b = &mce_banks[i];
+
                if (b->init)
                        wrmsrl(MSR_IA32_MCx_CTL(i), 0);
        }
@@ -1911,6 +1928,7 @@ static void mce_disable_cpu(void *h)
                cmci_clear();
        for (i = 0; i < banks; i++) {
                struct mce_bank *b = &mce_banks[i];
+
                if (b->init)
                        wrmsrl(MSR_IA32_MCx_CTL(i), 0);
        }
@@ -1928,6 +1946,7 @@ static void mce_reenable_cpu(void *h)
                cmci_reenable();
        for (i = 0; i < banks; i++) {
                struct mce_bank *b = &mce_banks[i];
+
                if (b->init)
                        wrmsrl(MSR_IA32_MCx_CTL(i), b->ctl);
        }
index 63a56d147e4a67b55ceea258dfeb20c18b9d1ccf..b3a1dba75330a4891b7d49739bd8507076a642da 100644 (file)
 /* How long to wait between reporting thermal events */
 #define CHECK_INTERVAL         (300 * HZ)
 
-static DEFINE_PER_CPU(__u64, next_check) = INITIAL_JIFFIES;
-static DEFINE_PER_CPU(unsigned long, thermal_throttle_count);
-static DEFINE_PER_CPU(bool, thermal_throttle_active);
+/*
+ * Current thermal throttling state:
+ */
+struct thermal_state {
+       bool                    is_throttled;
+
+       u64                     next_check;
+       unsigned long           throttle_count;
+       unsigned long           last_throttle_count;
+};
+
+static DEFINE_PER_CPU(struct thermal_state, thermal_state);
 
-static atomic_t therm_throt_en         = ATOMIC_INIT(0);
+static atomic_t therm_throt_en = ATOMIC_INIT(0);
 
 #ifdef CONFIG_SYSFS
 #define define_therm_throt_sysdev_one_ro(_name)                                \
        static SYSDEV_ATTR(_name, 0444, therm_throt_sysdev_show_##_name, NULL)
 
 #define define_therm_throt_sysdev_show_func(name)                      \
-static ssize_t therm_throt_sysdev_show_##name(struct sys_device *dev,  \
-                                       struct sysdev_attribute *attr,  \
-                                             char *buf)                \
+                                                                       \
+static ssize_t therm_throt_sysdev_show_##name(                         \
+                       struct sys_device *dev,                         \
+                       struct sysdev_attribute *attr,                  \
+                       char *buf)                                      \
 {                                                                      \
        unsigned int cpu = dev->id;                                     \
        ssize_t ret;                                                    \
@@ -55,7 +66,7 @@ static ssize_t therm_throt_sysdev_show_##name(struct sys_device *dev, \
        preempt_disable();      /* CPU hotplug */                       \
        if (cpu_online(cpu))                                            \
                ret = sprintf(buf, "%lu\n",                             \
-                             per_cpu(thermal_throttle_##name, cpu));   \
+                             per_cpu(thermal_state, cpu).name);        \
        else                                                            \
                ret = 0;                                                \
        preempt_enable();                                               \
@@ -63,11 +74,11 @@ static ssize_t therm_throt_sysdev_show_##name(struct sys_device *dev,       \
        return ret;                                                     \
 }
 
-define_therm_throt_sysdev_show_func(count);
-define_therm_throt_sysdev_one_ro(count);
+define_therm_throt_sysdev_show_func(throttle_count);
+define_therm_throt_sysdev_one_ro(throttle_count);
 
 static struct attribute *thermal_throttle_attrs[] = {
-       &attr_count.attr,
+       &attr_throttle_count.attr,
        NULL
 };
 
@@ -93,33 +104,39 @@ static struct attribute_group thermal_throttle_attr_group = {
  *          1 : Event should be logged further, and a message has been
  *              printed to the syslog.
  */
-static int therm_throt_process(int curr)
+static int therm_throt_process(bool is_throttled)
 {
-       unsigned int cpu = smp_processor_id();
-       __u64 tmp_jiffs = get_jiffies_64();
-       bool was_throttled = __get_cpu_var(thermal_throttle_active);
-       bool is_throttled = __get_cpu_var(thermal_throttle_active) = curr;
+       struct thermal_state *state;
+       unsigned int this_cpu;
+       bool was_throttled;
+       u64 now;
+
+       this_cpu = smp_processor_id();
+       now = get_jiffies_64();
+       state = &per_cpu(thermal_state, this_cpu);
+
+       was_throttled = state->is_throttled;
+       state->is_throttled = is_throttled;
 
        if (is_throttled)
-               __get_cpu_var(thermal_throttle_count)++;
+               state->throttle_count++;
 
-       if (!(was_throttled ^ is_throttled) &&
-           time_before64(tmp_jiffs, __get_cpu_var(next_check)))
+       if (time_before64(now, state->next_check) &&
+                       state->throttle_count != state->last_throttle_count)
                return 0;
 
-       __get_cpu_var(next_check) = tmp_jiffs + CHECK_INTERVAL;
+       state->next_check = now + CHECK_INTERVAL;
+       state->last_throttle_count = state->throttle_count;
 
        /* if we just entered the thermal event */
        if (is_throttled) {
-               printk(KERN_CRIT "CPU%d: Temperature above threshold, "
-                      "cpu clock throttled (total events = %lu)\n",
-                      cpu, __get_cpu_var(thermal_throttle_count));
+               printk(KERN_CRIT "CPU%d: Temperature above threshold, cpu clock throttled (total events = %lu)\n", this_cpu, state->throttle_count);
 
                add_taint(TAINT_MACHINE_CHECK);
                return 1;
        }
        if (was_throttled) {
-               printk(KERN_INFO "CPU%d: Temperature/speed normal\n", cpu);
+               printk(KERN_INFO "CPU%d: Temperature/speed normal\n", this_cpu);
                return 1;
        }
 
@@ -213,7 +230,7 @@ static void intel_thermal_interrupt(void)
        __u64 msr_val;
 
        rdmsrl(MSR_IA32_THERM_STATUS, msr_val);
-       if (therm_throt_process(msr_val & THERM_STATUS_PROCHOT))
+       if (therm_throt_process((msr_val & THERM_STATUS_PROCHOT) != 0))
                mce_log_therm_throt_event(msr_val);
 }
 
index 027b5b498993b6f0606367e2803f6fc9f15dca2c..f37930954d1596c8c366cfa87a5ddb629bb13513 100644 (file)
@@ -114,7 +114,7 @@ void __cpuinit check_tsc_sync_source(int cpu)
                return;
 
        if (boot_cpu_has(X86_FEATURE_TSC_RELIABLE)) {
-               pr_info("Skipping synchronization checks as TSC is reliable.\n");
+               printk_once(KERN_INFO "Skipping synchronization checks as TSC is reliable.\n");
                return;
        }
 
index 9b5a9f59a478781bd32025a938826e9c239e0d17..06630d26e56d1e01823901f525385eb002531bf6 100644 (file)
@@ -1,9 +1,10 @@
 obj-y  :=  init.o init_$(BITS).o fault.o ioremap.o extable.o pageattr.o mmap.o \
-           pat.o pgtable.o physaddr.o gup.o
+           pat.o pgtable.o physaddr.o gup.o setup_nx.o
 
 # Make sure __phys_addr has no stackprotector
 nostackp := $(call cc-option, -fno-stack-protector)
 CFLAGS_physaddr.o              := $(nostackp)
+CFLAGS_setup_nx.o              := $(nostackp)
 
 obj-$(CONFIG_SMP)              += tlb.o
 
index 0607119cef94f2321ead2e46d467d1900a7014fb..73ffd5536f62d1fcbdd7b0c3804415436ebe4dc1 100644 (file)
@@ -28,69 +28,6 @@ int direct_gbpages
 #endif
 ;
 
-int nx_enabled;
-
-#if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE)
-static int disable_nx __cpuinitdata;
-
-/*
- * noexec = on|off
- *
- * Control non-executable mappings for processes.
- *
- * on      Enable
- * off     Disable
- */
-static int __init noexec_setup(char *str)
-{
-       if (!str)
-               return -EINVAL;
-       if (!strncmp(str, "on", 2)) {
-               __supported_pte_mask |= _PAGE_NX;
-               disable_nx = 0;
-       } else if (!strncmp(str, "off", 3)) {
-               disable_nx = 1;
-               __supported_pte_mask &= ~_PAGE_NX;
-       }
-       return 0;
-}
-early_param("noexec", noexec_setup);
-#endif
-
-#ifdef CONFIG_X86_PAE
-static void __init set_nx(void)
-{
-       unsigned int v[4], l, h;
-
-       if (cpu_has_pae && (cpuid_eax(0x80000000) > 0x80000001)) {
-               cpuid(0x80000001, &v[0], &v[1], &v[2], &v[3]);
-
-               if ((v[3] & (1 << 20)) && !disable_nx) {
-                       rdmsr(MSR_EFER, l, h);
-                       l |= EFER_NX;
-                       wrmsr(MSR_EFER, l, h);
-                       nx_enabled = 1;
-                       __supported_pte_mask |= _PAGE_NX;
-               }
-       }
-}
-#else
-static inline void set_nx(void)
-{
-}
-#endif
-
-#ifdef CONFIG_X86_64
-void __cpuinit check_efer(void)
-{
-       unsigned long efer;
-
-       rdmsrl(MSR_EFER, efer);
-       if (!(efer & EFER_NX) || disable_nx)
-               __supported_pte_mask &= ~_PAGE_NX;
-}
-#endif
-
 static void __init find_early_table_space(unsigned long end, int use_pse,
                                          int use_gbpages)
 {
index 7257cf3decf9455ccb52f8f06bd28bc84a7f57ec..e78cd0ec2bcfa102d10af7ac7200af1506d00764 100644 (file)
@@ -81,6 +81,7 @@ enum {
 void pat_init(void)
 {
        u64 pat;
+       bool boot_cpu = !boot_pat_state;
 
        if (!pat_enabled)
                return;
@@ -122,8 +123,10 @@ void pat_init(void)
                rdmsrl(MSR_IA32_CR_PAT, boot_pat_state);
 
        wrmsrl(MSR_IA32_CR_PAT, pat);
-       printk(KERN_INFO "x86 PAT enabled: cpu %d, old 0x%Lx, new 0x%Lx\n",
-              smp_processor_id(), boot_pat_state, pat);
+
+       if (boot_cpu)
+               printk(KERN_INFO "x86 PAT enabled: cpu %d, old 0x%Lx, new 0x%Lx\n",
+                      smp_processor_id(), boot_pat_state, pat);
 }
 
 #undef PAT
diff --git a/arch/x86/mm/setup_nx.c b/arch/x86/mm/setup_nx.c
new file mode 100644 (file)
index 0000000..513d8ed
--- /dev/null
@@ -0,0 +1,69 @@
+#include <linux/spinlock.h>
+#include <linux/errno.h>
+#include <linux/init.h>
+
+#include <asm/pgtable.h>
+
+int nx_enabled;
+
+#if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE)
+static int disable_nx __cpuinitdata;
+
+/*
+ * noexec = on|off
+ *
+ * Control non-executable mappings for processes.
+ *
+ * on      Enable
+ * off     Disable
+ */
+static int __init noexec_setup(char *str)
+{
+       if (!str)
+               return -EINVAL;
+       if (!strncmp(str, "on", 2)) {
+               __supported_pte_mask |= _PAGE_NX;
+               disable_nx = 0;
+       } else if (!strncmp(str, "off", 3)) {
+               disable_nx = 1;
+               __supported_pte_mask &= ~_PAGE_NX;
+       }
+       return 0;
+}
+early_param("noexec", noexec_setup);
+#endif
+
+#ifdef CONFIG_X86_PAE
+void __init set_nx(void)
+{
+       unsigned int v[4], l, h;
+
+       if (cpu_has_pae && (cpuid_eax(0x80000000) > 0x80000001)) {
+               cpuid(0x80000001, &v[0], &v[1], &v[2], &v[3]);
+
+               if ((v[3] & (1 << 20)) && !disable_nx) {
+                       rdmsr(MSR_EFER, l, h);
+                       l |= EFER_NX;
+                       wrmsr(MSR_EFER, l, h);
+                       nx_enabled = 1;
+                       __supported_pte_mask |= _PAGE_NX;
+               }
+       }
+}
+#else
+void set_nx(void)
+{
+}
+#endif
+
+#ifdef CONFIG_X86_64
+void __cpuinit check_efer(void)
+{
+       unsigned long efer;
+
+       rdmsrl(MSR_EFER, efer);
+       if (!(efer & EFER_NX) || disable_nx)
+               __supported_pte_mask &= ~_PAGE_NX;
+}
+#endif
+
index 544eb7496531b43990d5f89d588832f0c8265814..3439616d69f188787a6b06f60cc901f37b69735c 100644 (file)
@@ -1082,6 +1082,11 @@ asmlinkage void __init xen_start_kernel(void)
 
        __supported_pte_mask |= _PAGE_IOMAP;
 
+#ifdef CONFIG_X86_64
+       /* Work out if we support NX */
+       check_efer();
+#endif
+
        xen_setup_features();
 
        /* Get mfn list */
@@ -1123,11 +1128,6 @@ asmlinkage void __init xen_start_kernel(void)
 
        pgd = (pgd_t *)xen_start_info->pt_base;
 
-#ifdef CONFIG_X86_64
-       /* Work out if we support NX */
-       check_efer();
-#endif
-
        /* Don't do the full vcpu_info placement stuff until we have a
           possible map and a non-dummy shared_info. */
        per_cpu(xen_vcpu, 0) = &HYPERVISOR_shared_info->vcpu_info[0];