3.10.70-rt75-rc1 stable review patch.
If anyone has any objections, please let me know.

------------------

From: Mike Galbraith <umgwanakikb...@gmail.com>

Shrug.  Lots of hobbyists have a beast in their basement, right?

Cc: stable...@vger.kernel.org
Signed-off-by: Mike Galbraith <mgalbra...@suse.de>
Signed-off-by: Sebastian Andrzej Siewior <bige...@linutronix.de>
Signed-off-by: Steven Rostedt <rost...@goodmis.org>
---
 arch/x86/include/asm/uv/uv_bau.h   | 14 +++++++-------
 arch/x86/include/asm/uv/uv_hub.h   |  2 +-
 arch/x86/kernel/apic/x2apic_uv_x.c |  2 +-
 arch/x86/platform/uv/tlb_uv.c      | 26 +++++++++++++-------------
 arch/x86/platform/uv/uv_time.c     | 21 +++++++++++++--------
 5 files changed, 35 insertions(+), 30 deletions(-)

diff --git a/arch/x86/include/asm/uv/uv_bau.h b/arch/x86/include/asm/uv/uv_bau.h
index a06983cdc125..afaaf2900796 100644
--- a/arch/x86/include/asm/uv/uv_bau.h
+++ b/arch/x86/include/asm/uv/uv_bau.h
@@ -611,9 +611,9 @@ struct bau_control {
        cycles_t                send_message;
        cycles_t                period_end;
        cycles_t                period_time;
-       spinlock_t              uvhub_lock;
-       spinlock_t              queue_lock;
-       spinlock_t              disable_lock;
+       raw_spinlock_t          uvhub_lock;
+       raw_spinlock_t          queue_lock;
+       raw_spinlock_t          disable_lock;
        /* tunables */
        int                     max_concurr;
        int                     max_concurr_const;
@@ -770,15 +770,15 @@ static inline int atom_asr(short i, struct atomic_short 
*v)
  * to be lowered below the current 'v'.  atomic_add_unless can only stop
  * on equal.
  */
-static inline int atomic_inc_unless_ge(spinlock_t *lock, atomic_t *v, int u)
+static inline int atomic_inc_unless_ge(raw_spinlock_t *lock, atomic_t *v, int 
u)
 {
-       spin_lock(lock);
+       raw_spin_lock(lock);
        if (atomic_read(v) >= u) {
-               spin_unlock(lock);
+               raw_spin_unlock(lock);
                return 0;
        }
        atomic_inc(v);
-       spin_unlock(lock);
+       raw_spin_unlock(lock);
        return 1;
 }
 
diff --git a/arch/x86/include/asm/uv/uv_hub.h b/arch/x86/include/asm/uv/uv_hub.h
index 2c32df95bb78..2c676e8d39ac 100644
--- a/arch/x86/include/asm/uv/uv_hub.h
+++ b/arch/x86/include/asm/uv/uv_hub.h
@@ -502,7 +502,7 @@ struct uv_blade_info {
        unsigned short  nr_online_cpus;
        unsigned short  pnode;
        short           memory_nid;
-       spinlock_t      nmi_lock;
+       raw_spinlock_t  nmi_lock;
        unsigned long   nmi_count;
 };
 extern struct uv_blade_info *uv_blade_info;
diff --git a/arch/x86/kernel/apic/x2apic_uv_x.c 
b/arch/x86/kernel/apic/x2apic_uv_x.c
index b32dbb411a9a..7a6ab4f38278 100644
--- a/arch/x86/kernel/apic/x2apic_uv_x.c
+++ b/arch/x86/kernel/apic/x2apic_uv_x.c
@@ -928,7 +928,7 @@ void __init uv_system_init(void)
                        uv_blade_info[blade].pnode = pnode;
                        uv_blade_info[blade].nr_possible_cpus = 0;
                        uv_blade_info[blade].nr_online_cpus = 0;
-                       spin_lock_init(&uv_blade_info[blade].nmi_lock);
+                       raw_spin_lock_init(&uv_blade_info[blade].nmi_lock);
                        min_pnode = min(pnode, min_pnode);
                        max_pnode = max(pnode, max_pnode);
                        blade++;
diff --git a/arch/x86/platform/uv/tlb_uv.c b/arch/x86/platform/uv/tlb_uv.c
index 0f92173a12b6..7ddc988cfa93 100644
--- a/arch/x86/platform/uv/tlb_uv.c
+++ b/arch/x86/platform/uv/tlb_uv.c
@@ -695,9 +695,9 @@ static void destination_plugged(struct bau_desc *bau_desc,
 
                quiesce_local_uvhub(hmaster);
 
-               spin_lock(&hmaster->queue_lock);
+               raw_spin_lock(&hmaster->queue_lock);
                reset_with_ipi(&bau_desc->distribution, bcp);
-               spin_unlock(&hmaster->queue_lock);
+               raw_spin_unlock(&hmaster->queue_lock);
 
                end_uvhub_quiesce(hmaster);
 
@@ -717,9 +717,9 @@ static void destination_timeout(struct bau_desc *bau_desc,
 
                quiesce_local_uvhub(hmaster);
 
-               spin_lock(&hmaster->queue_lock);
+               raw_spin_lock(&hmaster->queue_lock);
                reset_with_ipi(&bau_desc->distribution, bcp);
-               spin_unlock(&hmaster->queue_lock);
+               raw_spin_unlock(&hmaster->queue_lock);
 
                end_uvhub_quiesce(hmaster);
 
@@ -740,7 +740,7 @@ static void disable_for_period(struct bau_control *bcp, 
struct ptc_stats *stat)
        cycles_t tm1;
 
        hmaster = bcp->uvhub_master;
-       spin_lock(&hmaster->disable_lock);
+       raw_spin_lock(&hmaster->disable_lock);
        if (!bcp->baudisabled) {
                stat->s_bau_disabled++;
                tm1 = get_cycles();
@@ -753,7 +753,7 @@ static void disable_for_period(struct bau_control *bcp, 
struct ptc_stats *stat)
                        }
                }
        }
-       spin_unlock(&hmaster->disable_lock);
+       raw_spin_unlock(&hmaster->disable_lock);
 }
 
 static void count_max_concurr(int stat, struct bau_control *bcp,
@@ -816,7 +816,7 @@ static void record_send_stats(cycles_t time1, cycles_t 
time2,
  */
 static void uv1_throttle(struct bau_control *hmaster, struct ptc_stats *stat)
 {
-       spinlock_t *lock = &hmaster->uvhub_lock;
+       raw_spinlock_t *lock = &hmaster->uvhub_lock;
        atomic_t *v;
 
        v = &hmaster->active_descriptor_count;
@@ -948,7 +948,7 @@ static int check_enable(struct bau_control *bcp, struct 
ptc_stats *stat)
        struct bau_control *hmaster;
 
        hmaster = bcp->uvhub_master;
-       spin_lock(&hmaster->disable_lock);
+       raw_spin_lock(&hmaster->disable_lock);
        if (bcp->baudisabled && (get_cycles() >= bcp->set_bau_on_time)) {
                stat->s_bau_reenabled++;
                for_each_present_cpu(tcpu) {
@@ -960,10 +960,10 @@ static int check_enable(struct bau_control *bcp, struct 
ptc_stats *stat)
                                tbcp->period_giveups = 0;
                        }
                }
-               spin_unlock(&hmaster->disable_lock);
+               raw_spin_unlock(&hmaster->disable_lock);
                return 0;
        }
-       spin_unlock(&hmaster->disable_lock);
+       raw_spin_unlock(&hmaster->disable_lock);
        return -1;
 }
 
@@ -1880,9 +1880,9 @@ static void __init init_per_cpu_tunables(void)
                bcp->cong_reps                  = congested_reps;
                bcp->disabled_period =          sec_2_cycles(disabled_period);
                bcp->giveup_limit =             giveup_limit;
-               spin_lock_init(&bcp->queue_lock);
-               spin_lock_init(&bcp->uvhub_lock);
-               spin_lock_init(&bcp->disable_lock);
+               raw_spin_lock_init(&bcp->queue_lock);
+               raw_spin_lock_init(&bcp->uvhub_lock);
+               raw_spin_lock_init(&bcp->disable_lock);
        }
 }
 
diff --git a/arch/x86/platform/uv/uv_time.c b/arch/x86/platform/uv/uv_time.c
index 5c86786bbfd2..c039afa26aa2 100644
--- a/arch/x86/platform/uv/uv_time.c
+++ b/arch/x86/platform/uv/uv_time.c
@@ -58,7 +58,7 @@ static DEFINE_PER_CPU(struct clock_event_device, cpu_ced);
 
 /* There is one of these allocated per node */
 struct uv_rtc_timer_head {
-       spinlock_t      lock;
+       raw_spinlock_t  lock;
        /* next cpu waiting for timer, local node relative: */
        int             next_cpu;
        /* number of cpus on this node: */
@@ -178,7 +178,7 @@ static __init int uv_rtc_allocate_timers(void)
                                uv_rtc_deallocate_timers();
                                return -ENOMEM;
                        }
-                       spin_lock_init(&head->lock);
+                       raw_spin_lock_init(&head->lock);
                        head->ncpus = uv_blade_nr_possible_cpus(bid);
                        head->next_cpu = -1;
                        blade_info[bid] = head;
@@ -232,7 +232,7 @@ static int uv_rtc_set_timer(int cpu, u64 expires)
        unsigned long flags;
        int next_cpu;
 
-       spin_lock_irqsave(&head->lock, flags);
+       raw_spin_lock_irqsave(&head->lock, flags);
 
        next_cpu = head->next_cpu;
        *t = expires;
@@ -244,12 +244,12 @@ static int uv_rtc_set_timer(int cpu, u64 expires)
                if (uv_setup_intr(cpu, expires)) {
                        *t = ULLONG_MAX;
                        uv_rtc_find_next_timer(head, pnode);
-                       spin_unlock_irqrestore(&head->lock, flags);
+                       raw_spin_unlock_irqrestore(&head->lock, flags);
                        return -ETIME;
                }
        }
 
-       spin_unlock_irqrestore(&head->lock, flags);
+       raw_spin_unlock_irqrestore(&head->lock, flags);
        return 0;
 }
 
@@ -268,7 +268,7 @@ static int uv_rtc_unset_timer(int cpu, int force)
        unsigned long flags;
        int rc = 0;
 
-       spin_lock_irqsave(&head->lock, flags);
+       raw_spin_lock_irqsave(&head->lock, flags);
 
        if ((head->next_cpu == bcpu && uv_read_rtc(NULL) >= *t) || force)
                rc = 1;
@@ -280,7 +280,7 @@ static int uv_rtc_unset_timer(int cpu, int force)
                        uv_rtc_find_next_timer(head, pnode);
        }
 
-       spin_unlock_irqrestore(&head->lock, flags);
+       raw_spin_unlock_irqrestore(&head->lock, flags);
 
        return rc;
 }
@@ -300,13 +300,18 @@ static int uv_rtc_unset_timer(int cpu, int force)
 static cycle_t uv_read_rtc(struct clocksource *cs)
 {
        unsigned long offset;
+       cycle_t cycles;
 
+       preempt_disable();
        if (uv_get_min_hub_revision_id() == 1)
                offset = 0;
        else
                offset = (uv_blade_processor_id() * L1_CACHE_BYTES) % PAGE_SIZE;
 
-       return (cycle_t)uv_read_local_mmr(UVH_RTC | offset);
+       cycles = (cycle_t)uv_read_local_mmr(UVH_RTC | offset);
+       preempt_enable();
+
+       return cycles;
 }
 
 /*
-- 
2.1.4


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to