fixes trailing whitespace and spaces before tab indents in 2.6.20-rc5-rt7 as
reported with: git-apply --whitespace=error-all

---

 arch/arm/mach-omap1/time.c        |    2 +-
 arch/i386/kernel/entry.S          |    2 +-
 arch/i386/kernel/reboot.c         |    2 +-
 arch/ia64/Kconfig                 |    4 ++--
 arch/ia64/kernel/fsys.S           |    2 +-
 arch/ia64/kernel/time.c           |    4 ++--
 arch/x86_64/kernel/entry.S        |    2 +-
 arch/x86_64/kernel/hpet.c         |    2 +-
 arch/x86_64/kernel/reboot.c       |    2 +-
 arch/x86_64/kernel/smpboot.c      |    8 ++++----
 arch/x86_64/kernel/tsc.c          |    8 ++++----
 arch/x86_64/kernel/vsyscall.c     |    2 +-
 drivers/char/lpptest.c            |    2 +-
 include/asm-arm/atomic.h          |    2 +-
 include/asm-generic/vmlinux.lds.h |    4 ++--
 include/linux/irq.h               |    6 +++---
 include/linux/mutex.h             |    2 +-
 kernel/fork.c                     |    6 +++---
 kernel/futex.c                    |    4 ++--
 kernel/hrtimer.c                  |    2 +-
 kernel/latency_trace.c            |    6 +++---
 kernel/spinlock.c                 |    2 +-
 kernel/workqueue.c                |    2 +-
 mm/migrate.c                      |    2 +-
 mm/slab.c                         |   14 +++++++-------
 25 files changed, 47 insertions(+), 47 deletions(-)

diff --git a/arch/arm/mach-omap1/time.c b/arch/arm/mach-omap1/time.c
index 7baf0df..85f9f5e 100644
--- a/arch/arm/mach-omap1/time.c
+++ b/arch/arm/mach-omap1/time.c
@@ -105,7 +105,7 @@ static inline unsigned long omap_mpu_tim
 
 static inline void omap_mpu_set_autoreset(int nr)
 {
-       volatile omap_mpu_timer_regs_t* timer = omap_mpu_timer_base(nr);
+       volatile omap_mpu_timer_regs_t* timer = omap_mpu_timer_base(nr);
 
        timer->cntl = timer->cntl | MPU_TIMER_AR;
 }
diff --git a/arch/i386/kernel/entry.S b/arch/i386/kernel/entry.S
index ce8a092..33d6b80 100644
--- a/arch/i386/kernel/entry.S
+++ b/arch/i386/kernel/entry.S
@@ -272,7 +272,7 @@ need_resched:
        jz restore_nocheck
        testl $IF_MASK,PT_EFLAGS(%esp)  # interrupts off (exception path) ?
        jz restore_nocheck
-       DISABLE_INTERRUPTS(CLBR_ANY)
+       DISABLE_INTERRUPTS(CLBR_ANY)
 
        call preempt_schedule_irq
        jmp need_resched
diff --git a/arch/i386/kernel/reboot.c b/arch/i386/kernel/reboot.c
index 5f81e80..a3e7410 100644
--- a/arch/i386/kernel/reboot.c
+++ b/arch/i386/kernel/reboot.c
@@ -327,7 +327,7 @@ void machine_emergency_restart(void)
                asm volatile (
                        "1: .byte 0x0f, 0x01, 0xc4      \n" /* vmxoff */
                        "2:                             \n"
-                       ".section __ex_table,\"a\"      \n"
+                       ".section __ex_table,\"a\"      \n"
                        "   .align 4                    \n"
                        "   .long       1b,2b           \n"
                        ".previous                      \n"
diff --git a/arch/ia64/Kconfig b/arch/ia64/Kconfig
index 333049c..a1a6bc3 100644
--- a/arch/ia64/Kconfig
+++ b/arch/ia64/Kconfig
@@ -300,7 +300,7 @@ config HIGH_RES_RESOLUTION
 choice
        prompt "Clock source"
        depends on HIGH_RES_TIMERS
-       default HIGH_RES_TIMER_ITC
+       default HIGH_RES_TIMER_ITC
        help
          This option allows you to choose the hardware source in charge
          of generating high precision interruptions on your system.
@@ -308,7 +308,7 @@ choice
 
          <timer>                               <resolution>
          ITC Interval Time Counter             1/CPU clock
-         HPET High Precision Event Timer       ~ (XXX:have to check the spec)
+         HPET High Precision Event Timer       ~ (XXX:have to check the spec)
 
          The ITC timer is available on all the ia64 computers because
          it is integrated directly into the processor. However it may not
diff --git a/arch/ia64/kernel/fsys.S b/arch/ia64/kernel/fsys.S
index 7b94c34..2abf920 100644
--- a/arch/ia64/kernel/fsys.S
+++ b/arch/ia64/kernel/fsys.S
@@ -219,7 +219,7 @@ ENTRY(fsys_gettimeofday)
 (p6)    br.cond.spnt.few .fail_einval  // deferred branch
        ;;
        ld8 r30 = [r10]         // clocksource->mmio_ptr
-       movl r19 = itc_lastcycle 
+       movl r19 = itc_lastcycle
        add r23 = IA64_CLOCKSOURCE_SHIFT_OFFSET,r20
        cmp.ne p6, p0 = 0, r2   // Fallback if work is scheduled
 (p6)    br.cond.spnt.many fsys_fallback_syscall
diff --git a/arch/ia64/kernel/time.c b/arch/ia64/kernel/time.c
index 983ef26..4fc3670 100644
--- a/arch/ia64/kernel/time.c
+++ b/arch/ia64/kernel/time.c
@@ -262,10 +262,10 @@ ia64_init_itm (void)
        ia64_cpu_local_tick();
 
        if (!clocksource_itc_p) {
-               /* Sort out mult/shift values: */
+               /* Sort out mult/shift values: */
                clocksource_itc.mult = 
clocksource_hz2mult(local_cpu_data->itc_freq,
                                                   clocksource_itc.shift);
-               clocksource_register(&clocksource_itc);
+               clocksource_register(&clocksource_itc);
                clocksource_itc_p = &clocksource_itc;
        }
 }
diff --git a/arch/x86_64/kernel/entry.S b/arch/x86_64/kernel/entry.S
index fb0dace..75f2626 100644
--- a/arch/x86_64/kernel/entry.S
+++ b/arch/x86_64/kernel/entry.S
@@ -361,7 +361,7 @@ tracesys:
        movq %r10,%rcx  /* fixup for C */
        TRACE_SYS_CALL
        call *sys_call_table(,%rax,8)
-       TRACE_SYS_RET
+       TRACE_SYS_RET
 1:     movq %rax,RAX-ARGOFFSET(%rsp)
        /* Use IRET because user could have changed frame */
                
diff --git a/arch/x86_64/kernel/hpet.c b/arch/x86_64/kernel/hpet.c
index 9edc2e4..8ae86ff 100644
--- a/arch/x86_64/kernel/hpet.c
+++ b/arch/x86_64/kernel/hpet.c
@@ -63,7 +63,7 @@ static __init int late_hpet_init(void)
        unsigned int            ntimer;
 
        if (!hpet_address)
-               return 0;
+               return 0;
 
        memset(&hd, 0, sizeof (hd));
 
diff --git a/arch/x86_64/kernel/reboot.c b/arch/x86_64/kernel/reboot.c
index dda0b88..36bad04 100644
--- a/arch/x86_64/kernel/reboot.c
+++ b/arch/x86_64/kernel/reboot.c
@@ -124,7 +124,7 @@ void machine_emergency_restart(void)
                asm volatile (
                        "1: .byte 0x0f, 0x01, 0xc4      \n" /* vmxoff */
                        "2:                             \n"
-                       ".section __ex_table,\"a\"      \n"
+                       ".section __ex_table,\"a\"      \n"
                        "   .align 8                    \n"
                        "   .quad       1b,2b           \n"
                        ".previous                      \n"
diff --git a/arch/x86_64/kernel/smpboot.c b/arch/x86_64/kernel/smpboot.c
index f2c0475..8551845 100644
--- a/arch/x86_64/kernel/smpboot.c
+++ b/arch/x86_64/kernel/smpboot.c
@@ -335,8 +335,8 @@ void __cpuinit notrace start_secondary(v
        barrier();
 
        /*
-        * Check TSC sync first:
-        */
+        * Check TSC sync first:
+        */
        check_tsc_sync_target();
 
        Dprintk("cpu %d: setting up apic clock\n", smp_processor_id());         
@@ -954,8 +954,8 @@ int __cpuinit __cpu_up(unsigned int cpu)
        Dprintk("waiting for cpu %d\n", cpu);
 
        /*
-        * Make sure and check TSC sync:
-        */
+        * Make sure and check TSC sync:
+        */
        check_tsc_sync_source(cpu);
 
        while (!cpu_isset(cpu, cpu_online_map))
diff --git a/arch/x86_64/kernel/tsc.c b/arch/x86_64/kernel/tsc.c
index ef2ec93..655d170 100644
--- a/arch/x86_64/kernel/tsc.c
+++ b/arch/x86_64/kernel/tsc.c
@@ -157,17 +157,17 @@ __cpuinit int unsynchronized_tsc(void)
 #endif
        /* Most intel systems have synchronized TSCs except for
           multi node systems */
-       if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) {
+       if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) {
 #ifdef CONFIG_ACPI
                /* But TSC doesn't tick in C3 so don't use it there */
                if (acpi_fadt.length > 0 && acpi_fadt.plvl3_lat < 100)
                        return 1;
 #endif
-               return 0;
+               return 0;
        }
 
-       /* Assume multi socket systems are not synchronized */
-       return num_present_cpus() > 1;
+       /* Assume multi socket systems are not synchronized */
+       return num_present_cpus() > 1;
 }
 
 int __init notsc_setup(char *s)
diff --git a/arch/x86_64/kernel/vsyscall.c b/arch/x86_64/kernel/vsyscall.c
index b681eee..ed7c882 100644
--- a/arch/x86_64/kernel/vsyscall.c
+++ b/arch/x86_64/kernel/vsyscall.c
@@ -73,7 +73,7 @@ void update_vsyscall(struct timespec* wa
        vsyscall_gtod_data.wall_time_tv.tv_usec = wall_time->tv_nsec/1000;
        vsyscall_gtod_data.sys_tz = sys_tz;
 
-       if (cpu_has(&boot_cpu_data, X86_FEATURE_RDTSCP))
+       if (cpu_has(&boot_cpu_data, X86_FEATURE_RDTSCP))
                vsyscall_gtod_data.vgetcpu_mode = VGETCPU_RDTSCP;
        else
                vsyscall_gtod_data.vgetcpu_mode = VGETCPU_LSL;
diff --git a/drivers/char/lpptest.c b/drivers/char/lpptest.c
index 02765fe..0c9116b 100644
--- a/drivers/char/lpptest.c
+++ b/drivers/char/lpptest.c
@@ -78,7 +78,7 @@ static cycles_t test_response(void)
        outb(0x08, 0x378);
        now = get_cycles();
        while(1) {
-               if (inb(0x379) != in)
+               if (inb(0x379) != in)
                        break;
                if (timeout++ > 1000000) {
                        outb(0x00, 0x378);
diff --git a/include/asm-arm/atomic.h b/include/asm-arm/atomic.h
index e21494d..9846347 100644
--- a/include/asm-arm/atomic.h
+++ b/include/asm-arm/atomic.h
@@ -242,7 +242,7 @@ static inline unsigned long __cmpxchg(vo
      __typeof__(*(ptr)) _o_ = (o);                                     \
      __typeof__(*(ptr)) _n_ = (n);                                     \
      (__typeof__(*(ptr))) __cmpxchg((ptr), (unsigned long)_o_,         \
-                                (unsigned long)_n_, sizeof(*(ptr)));   \
+                               (unsigned long)_n_, sizeof(*(ptr)));    \
 })
 
 #endif
diff --git a/include/asm-generic/vmlinux.lds.h 
b/include/asm-generic/vmlinux.lds.h
index 4c595cb..8ebba62 100644
--- a/include/asm-generic/vmlinux.lds.h
+++ b/include/asm-generic/vmlinux.lds.h
@@ -167,7 +167,7 @@
 #define EH_FRAME                                                       \
                /* Unwind data binary search table */                   \
                . = ALIGN(8);                                           \
-               .eh_frame_hdr : AT(ADDR(.eh_frame_hdr) - LOAD_OFFSET) { \
+               .eh_frame_hdr : AT(ADDR(.eh_frame_hdr) - LOAD_OFFSET) { \
                        VMLINUX_SYMBOL(__start_unwind_hdr) = .;         \
                        *(.eh_frame_hdr)                                \
                        VMLINUX_SYMBOL(__end_unwind_hdr) = .;           \
@@ -176,7 +176,7 @@
                . = ALIGN(8);                                           \
                .eh_frame : AT(ADDR(.eh_frame) - LOAD_OFFSET) {         \
                        VMLINUX_SYMBOL(__start_unwind) = .;             \
-                       *(.eh_frame)                                    \
+                       *(.eh_frame)                                    \
                        VMLINUX_SYMBOL(__end_unwind) = .;               \
                }
 #else
diff --git a/include/linux/irq.h b/include/linux/irq.h
index 28f049a..0f3bc01 100644
--- a/include/linux/irq.h
+++ b/include/linux/irq.h
@@ -161,9 +161,9 @@ struct irq_desc {
        unsigned int            wake_depth;     /* nested wake enables */
        unsigned int            irq_count;      /* For detecting broken IRQs */
        unsigned int            irqs_unhandled;
-       struct task_struct      *thread;
-       wait_queue_head_t       wait_for_handler;
-       cycles_t                timestamp;
+       struct task_struct      *thread;
+       wait_queue_head_t       wait_for_handler;
+       cycles_t                timestamp;
        raw_spinlock_t          lock;
 #ifdef CONFIG_SMP
        cpumask_t               affinity;
diff --git a/include/linux/mutex.h b/include/linux/mutex.h
index 7a5d1fe..25d9df8 100644
--- a/include/linux/mutex.h
+++ b/include/linux/mutex.h
@@ -30,7 +30,7 @@ struct mutex {
 };
 
 #define DEFINE_MUTEX(mutexname)                                        \
-       struct mutex mutexname = {                                      \
+       struct mutex mutexname = {                                      \
                .lock = __RT_MUTEX_INITIALIZER(mutexname.lock)          \
        }
 
diff --git a/kernel/fork.c b/kernel/fork.c
index 075067e..1630673 100644
--- a/kernel/fork.c
+++ b/kernel/fork.c
@@ -1876,9 +1876,9 @@ static int __devinit cpu_callback(struct
                        printk("desched_thread for %i failed\n", hotcpu);
                        return NOTIFY_BAD;
                }
-               per_cpu(desched_task, hotcpu) = p;
+               per_cpu(desched_task, hotcpu) = p;
                kthread_bind(p, hotcpu);
-               break;
+               break;
        case CPU_ONLINE:
 
                wake_up_process(per_cpu(desched_task, hotcpu));
@@ -1896,7 +1896,7 @@ static int __devinit cpu_callback(struct
                takeover_tasklets(hotcpu);
                break;
 #endif /* CONFIG_HOTPLUG_CPU */
-       }
+       }
        return NOTIFY_OK;
 }
 
diff --git a/kernel/futex.c b/kernel/futex.c
index b7ca553..9507575 100644
--- a/kernel/futex.c
+++ b/kernel/futex.c
@@ -1358,7 +1358,7 @@ static int fixup_pi_state_owner(u32 __us
                curval = futex_atomic_cmpxchg_inatomic(uaddr,
                                                       uval, newval);
                if (curval == -EFAULT)
-                       ret = -EFAULT;
+                       ret = -EFAULT;
                if (curval == uval)
                        break;
                uval = curval;
@@ -2380,7 +2380,7 @@ static int fixup_pi_state_owner64(u64 __
                curval = futex_atomic_cmpxchg_inatomic64(uaddr,
                                                         uval, newval);
                if (curval == -EFAULT)
-                       ret = -EFAULT;
+                       ret = -EFAULT;
                if (curval == uval)
                        break;
                uval = curval;
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c
index fb7731f..dfa9d0f 100644
--- a/kernel/hrtimer.c
+++ b/kernel/hrtimer.c
@@ -767,7 +767,7 @@ static enum hrtimer_restart hrtimer_sche
 {
        struct hrtimer_cpu_base *cpu_base =
                container_of(timer, struct hrtimer_cpu_base, sched_timer);
-       struct pt_regs *regs = get_irq_regs();
+       struct pt_regs *regs = get_irq_regs();
        ktime_t now = ktime_get();
 
        /* Check, if the jiffies need an update */
diff --git a/kernel/latency_trace.c b/kernel/latency_trace.c
index 936238a..7f6c304 100644
--- a/kernel/latency_trace.c
+++ b/kernel/latency_trace.c
@@ -546,7 +546,7 @@ static void notrace early_print_entry(st
                (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
                (entry->flags & TRACE_FLAG_IRQS_HARD_OFF) ? 'D' : '.',
                (entry->flags & TRACE_FLAG_NEED_RESCHED_DELAYED) ? 'n' :
-               ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
+               ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
 
        hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
        softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
@@ -1321,7 +1321,7 @@ print_generic(struct seq_file *m, struct
                (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
                (entry->flags & TRACE_FLAG_IRQS_HARD_OFF) ? 'D' : '.',
                (entry->flags & TRACE_FLAG_NEED_RESCHED_DELAYED) ? 'n' :
-               ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
+               ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
 
        hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
        softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
@@ -2483,7 +2483,7 @@ static void print_entry(struct trace_ent
                (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
                (entry->flags & TRACE_FLAG_IRQS_HARD_OFF) ? 'D' : '.',
                (entry->flags & TRACE_FLAG_NEED_RESCHED_DELAYED) ? 'n' :
-               ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
+               ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
 
        hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
        softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
diff --git a/kernel/spinlock.c b/kernel/spinlock.c
index 95a4b60..2b0b792 100644
--- a/kernel/spinlock.c
+++ b/kernel/spinlock.c
@@ -252,7 +252,7 @@ void __lockfunc __##op##_lock(locktype##
                if (!(lock)->break_lock)                                \
                        (lock)->break_lock = 1;                         \
                while (!__raw_##op##_can_lock(&(lock)->raw_lock) &&     \
-                                       (lock)->break_lock)             \
+                                       (lock)->break_lock)             \
                        __raw_##op##_relax(&lock->raw_lock);            \
        }                                                               \
        (lock)->break_lock = 0;                                         \
diff --git a/kernel/workqueue.c b/kernel/workqueue.c
index b532f38..e685d6d 100644
--- a/kernel/workqueue.c
+++ b/kernel/workqueue.c
@@ -309,7 +309,7 @@ static void leak_check(void *func)
        printk(KERN_ERR "BUG: workqueue leaked lock or atomic: "
                                "%s/0x%08x/%d\n",
                                current->comm, preempt_count(),
-                               current->pid);
+                               current->pid);
        printk(KERN_ERR "    last function: ");
        print_symbol("%s\n", (unsigned long)func);
        debug_show_held_locks(current);
diff --git a/mm/migrate.c b/mm/migrate.c
index e35e95e..7bf090c 100644
--- a/mm/migrate.c
+++ b/mm/migrate.c
@@ -329,7 +329,7 @@ static int migrate_page_move_mapping(str
 
        radix_tree_replace_slot(pslot, newpage);
        page->mapping = NULL;
-       write_unlock_irq(&mapping->tree_lock);
+       write_unlock_irq(&mapping->tree_lock);
        end_page_no_new_refs(page);
 
        /*
diff --git a/mm/slab.c b/mm/slab.c
index 038812c..bec4763 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -143,13 +143,13 @@
 # define slab_irq_disable_rt(flags)    do { (void)(flags); } while (0)
 # define slab_irq_enable_rt(flags)     do { (void)(flags); } while (0)
 # define slab_spin_lock_irq(lock, cpu) \
-       do { spin_lock_irq(lock); (cpu) = smp_processor_id(); } while (0)
+       do { spin_lock_irq(lock); (cpu) = smp_processor_id(); } while (0)
 # define slab_spin_unlock_irq(lock, cpu) \
-                                       spin_unlock_irq(lock)
+                                       spin_unlock_irq(lock)
 # define slab_spin_lock_irqsave(lock, flags, cpu) \
-       do { spin_lock_irqsave(lock, flags); (cpu) = smp_processor_id(); } 
while (0)
+       do { spin_lock_irqsave(lock, flags); (cpu) = smp_processor_id(); } 
while (0)
 # define slab_spin_unlock_irqrestore(lock, flags, cpu) \
-       do { spin_unlock_irqrestore(lock, flags); } while (0)
+       do { spin_unlock_irqrestore(lock, flags); } while (0)
 #else
 DEFINE_PER_CPU_LOCKED(int, slab_irq_locks) = { 0, };
 # define slab_irq_disable(cpu)         
(void)get_cpu_var_locked(slab_irq_locks, &(cpu))
@@ -167,9 +167,9 @@ DEFINE_PER_CPU_LOCKED(int, slab_irq_lock
 # define slab_spin_unlock_irq(lock, cpu) \
                do { spin_unlock(lock); slab_irq_enable(cpu); } while (0)
 # define slab_spin_lock_irqsave(lock, flags, cpu) \
-       do { slab_irq_disable(cpu); spin_lock_irqsave(lock, flags); } while (0)
+       do { slab_irq_disable(cpu); spin_lock_irqsave(lock, flags); } while (0)
 # define slab_spin_unlock_irqrestore(lock, flags, cpu) \
-       do { spin_unlock_irqrestore(lock, flags); slab_irq_enable(cpu); } while 
(0)
+       do { spin_unlock_irqrestore(lock, flags); slab_irq_enable(cpu); } while 
(0)
 #endif
 
 /*
@@ -3335,7 +3335,7 @@ static __always_inline void *__cache_all
         * ____cache_alloc_node() knows how to locate memory on other nodes
         */
        if (NUMA_BUILD && !objp)
-               objp = ____cache_alloc_node(cachep, flags, 
cpu_to_node(this_cpu), &this_cpu);
+               objp = ____cache_alloc_node(cachep, flags, 
cpu_to_node(this_cpu), &this_cpu);
        slab_irq_restore(irqflags, this_cpu);
        objp = cache_alloc_debugcheck_after(cachep, flags, objp,
                                            caller);
-- 
1.4.4.1

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to