diff -ruNp 500-version-specific-i386.patch-old/arch/i386/power/Makefile 500-version-specific-i386.patch-new/arch/i386/power/Makefile --- 500-version-specific-i386.patch-old/arch/i386/power/Makefile 2004-11-03 21:52:57.000000000 +1100 +++ 500-version-specific-i386.patch-new/arch/i386/power/Makefile 2005-07-04 23:14:19.000000000 +1000 @@ -1,2 +1,2 @@ -obj-$(CONFIG_PM) += cpu.o +obj-$(CONFIG_PM) += cpu.o smp.o obj-$(CONFIG_SOFTWARE_SUSPEND) += swsusp.o diff -ruNp 500-version-specific-i386.patch-old/arch/i386/power/smp.c 500-version-specific-i386.patch-new/arch/i386/power/smp.c --- 500-version-specific-i386.patch-old/arch/i386/power/smp.c 1970-01-01 10:00:00.000000000 +1000 +++ 500-version-specific-i386.patch-new/arch/i386/power/smp.c 2005-07-04 23:14:19.000000000 +1000 @@ -0,0 +1,84 @@ +//#ifdef CONFIG_SMP +#include <linux/suspend.h> +#include <linux/irq.h> +#include <asm/tlbflush.h> +#include <asm/suspend2.h> + +extern atomic_t suspend_cpu_counter __nosavedata; +unsigned char * my_saved_context __nosavedata; +static unsigned long c_loops_per_jiffy_ref[NR_CPUS] __nosavedata; + +struct suspend2_saved_context suspend2_saved_contexts[NR_CPUS]; +struct suspend2_saved_context suspend2_saved_context; /* temporary storage */ +cpumask_t saved_affinity[NR_IRQS]; + +/* + * Save and restore processor state for secondary processors. + * IRQs (and therefore preemption) are already disabled + * when we enter here (IPI). + */ + +static volatile int loop __nosavedata; + +void __smp_suspend_lowlevel(void * data) +{ + __asm__( "movl %%ecx,%%cr3\n" ::"c"(__pa(swsusp_pg_dir))); + + if (test_suspend_state(SUSPEND_NOW_RESUMING)) { + BUG_ON(!irqs_disabled()); + kernel_fpu_begin(); + c_loops_per_jiffy_ref[_smp_processor_id()] = current_cpu_data.loops_per_jiffy; + atomic_inc(&suspend_cpu_counter); + + /* Only image copied back while we spin in this loop. Our + * task info should not be looked at while this is happening + * (which smp_processor_id() will do( */ + while (test_suspend_state(SUSPEND_FREEZE_SMP)) { + cpu_relax(); + barrier(); + } + + while (atomic_read(&suspend_cpu_counter) != _smp_processor_id()) { + cpu_relax(); + barrier(); + } + my_saved_context = (unsigned char *) (suspend2_saved_contexts + _smp_processor_id()); + for (loop = sizeof(struct suspend2_saved_context); loop--; loop) + *(((unsigned char *) &suspend2_saved_context) + loop - 1) = *(my_saved_context + loop - 1); + suspend2_restore_processor_context(); + cpu_clear(_smp_processor_id(), per_cpu(cpu_tlbstate, _smp_processor_id()).active_mm->cpu_vm_mask); + load_cr3(swapper_pg_dir); + wbinvd(); + __flush_tlb_all(); + current_cpu_data.loops_per_jiffy = c_loops_per_jiffy_ref[_smp_processor_id()]; + mtrr_restore_one_cpu(); + atomic_dec(&suspend_cpu_counter); + } else { /* suspending */ + BUG_ON(!irqs_disabled()); + /* + *Save context and go back to idling. + * Note that we cannot leave the processor + * here. It must be able to receive IPIs if + * the LZF compression driver (eg) does a + * vfree after compressing the kernel etc + */ + while (test_suspend_state(SUSPEND_FREEZE_SMP) && + (atomic_read(&suspend_cpu_counter) != (_smp_processor_id() - 1))) { + cpu_relax(); + barrier(); + } + suspend2_save_processor_context(); + my_saved_context = (unsigned char *) (suspend2_saved_contexts + _smp_processor_id()); + for (loop = sizeof(struct suspend2_saved_context); loop--; loop) + *(my_saved_context + loop - 1) = *(((unsigned char *) &suspend2_saved_context) + loop - 1); + atomic_inc(&suspend_cpu_counter); + /* Now spin until the atomic copy of the kernel is made. */ + while (test_suspend_state(SUSPEND_FREEZE_SMP)) { + cpu_relax(); + barrier(); + } + atomic_dec(&suspend_cpu_counter); + kernel_fpu_end(); + } +} +//#endif diff -ruNp 500-version-specific-i386.patch-old/include/asm-i386/suspend2.h 500-version-specific-i386.patch-new/include/asm-i386/suspend2.h --- 500-version-specific-i386.patch-old/include/asm-i386/suspend2.h 1970-01-01 10:00:00.000000000 +1000 +++ 500-version-specific-i386.patch-new/include/asm-i386/suspend2.h 2005-07-05 23:56:41.000000000 +1000 @@ -0,0 +1,391 @@ + /* + * Copyright 2003-2005 Nigel Cunningham <[EMAIL PROTECTED]> + * Based on code + * Copyright 2001-2002 Pavel Machek <[EMAIL PROTECTED]> + * Based on code + * Copyright 2001 Patrick Mochel <[EMAIL PROTECTED]> + */ +#include <linux/irq.h> +#include <asm/desc.h> +#include <asm/i387.h> +#include <asm/tlbflush.h> +#include <asm/desc.h> +#include <asm/suspend.h> +#undef inline +#define inline __inline__ __attribute__((always_inline)) + +/* image of the saved processor states */ +struct suspend2_saved_context { + u32 eax, ebx, ecx, edx; + u32 esp, ebp, esi, edi; + u16 es, fs, gs, ss; + u32 cr0, cr2, cr3, cr4; + u16 gdt_pad; + u16 gdt_limit; + u32 gdt_base; + u16 idt_pad; + u16 idt_limit; + u32 idt_base; + u16 ldt; + u16 tss; + u32 tr; + u32 safety; + u32 return_address; + u32 eflags; +} __attribute__((packed)); + +extern struct suspend2_saved_context suspend2_saved_context; /* temporary storage */ + +#ifdef CONFIG_MTRR +/* MTRR functions */ +extern int mtrr_save(void); +extern int mtrr_restore_one_cpu(void); +extern void mtrr_restore_finish(void); +#else +#define mtrr_save() do { } while(0) +#define mtrr_restore_one_cpu() do { } while(0) +#define mtrr_restore_finish() do { } while(0) +#endif + +#ifndef CONFIG_SMP +#undef cpu_clear +#define cpu_clear(a, b) do { } while(0) +#endif + +extern struct suspend2_saved_context suspend2_saved_context; /* temporary storage */ + +/* + * save_processor_context + * + * Save the state of the processor before we go to sleep. + * + * return_stack is the value of the stack pointer (%esp) as the caller sees it. + * A good way could not be found to obtain it from here (don't want to make _too_ + * many assumptions about the layout of the stack this far down.) Also, the + * handy little __builtin_frame_pointer(level) where level > 0, is blatantly + * buggy - it returns the value of the stack at the proper location, not the + * location, like it should (as of gcc 2.91.66) + * + * Note that the context and timing of this function is pretty critical. + * With a minimal amount of things going on in the caller and in here, gcc + * does a good job of being just a dumb compiler. Watch the assembly output + * if anything changes, though, and make sure everything is going in the right + * place. + */ +static inline void suspend2_save_processor_context(void) +{ + kernel_fpu_begin(); + + /* + * descriptor tables + */ + asm volatile ("sgdt (%0)" : "=m" (suspend2_saved_context.gdt_limit)); + asm volatile ("sidt (%0)" : "=m" (suspend2_saved_context.idt_limit)); + asm volatile ("sldt (%0)" : "=m" (suspend2_saved_context.ldt)); + asm volatile ("str (%0)" : "=m" (suspend2_saved_context.tr)); + + /* + * save the general registers. + * note that gcc has constructs to specify output of certain registers, + * but they're not used here, because it assumes that you want to modify + * those registers, so it tries to be smart and save them beforehand. + * It's really not necessary, and kinda fishy (check the assembly output), + * so it's avoided. + */ + asm volatile ("movl %%esp, (%0)" : "=m" (suspend2_saved_context.esp)); + asm volatile ("movl %%eax, (%0)" : "=m" (suspend2_saved_context.eax)); + asm volatile ("movl %%ebx, (%0)" : "=m" (suspend2_saved_context.ebx)); + asm volatile ("movl %%ecx, (%0)" : "=m" (suspend2_saved_context.ecx)); + asm volatile ("movl %%edx, (%0)" : "=m" (suspend2_saved_context.edx)); + asm volatile ("movl %%ebp, (%0)" : "=m" (suspend2_saved_context.ebp)); + asm volatile ("movl %%esi, (%0)" : "=m" (suspend2_saved_context.esi)); + asm volatile ("movl %%edi, (%0)" : "=m" (suspend2_saved_context.edi)); + + /* + * segment registers + */ + asm volatile ("movw %%es, %0" : "=r" (suspend2_saved_context.es)); + asm volatile ("movw %%fs, %0" : "=r" (suspend2_saved_context.fs)); + asm volatile ("movw %%gs, %0" : "=r" (suspend2_saved_context.gs)); + asm volatile ("movw %%ss, %0" : "=r" (suspend2_saved_context.ss)); + + /* + * control registers + */ + asm volatile ("movl %%cr0, %0" : "=r" (suspend2_saved_context.cr0)); + asm volatile ("movl %%cr2, %0" : "=r" (suspend2_saved_context.cr2)); + asm volatile ("movl %%cr3, %0" : "=r" (suspend2_saved_context.cr3)); + asm volatile ("movl %%cr4, %0" : "=r" (suspend2_saved_context.cr4)); + + /* + * eflags + */ + asm volatile ("pushfl ; popl (%0)" : "=m" (suspend2_saved_context.eflags)); +} + +static void fix_processor_context(void) +{ + int nr = _smp_processor_id(); + struct tss_struct * t = &per_cpu(init_tss,nr); + + set_tss_desc(nr,t); /* This just modifies memory; should not be neccessary. But... This is neccessary, because 386 hardware has concept of busy tsc or some similar stupidity. */ + per_cpu(cpu_gdt_table,nr)[GDT_ENTRY_TSS].b &= 0xfffffdff; + + load_TR_desc(); + + load_LDT(¤t->active_mm->context); /* This does lldt */ + + /* + * Now maybe reload the debug registers + */ + if (current->thread.debugreg[7]){ + loaddebug(¤t->thread, 0); + loaddebug(¤t->thread, 1); + loaddebug(¤t->thread, 2); + loaddebug(¤t->thread, 3); + /* no 4 and 5 */ + loaddebug(¤t->thread, 6); + loaddebug(¤t->thread, 7); + } + +} + +static void do_fpu_end(void) +{ + /* restore FPU regs if necessary */ + /* Do it out of line so that gcc does not move cr0 load to some stupid place */ + kernel_fpu_end(); +} + +/* + * restore_processor_context + * + * Restore the processor context as it was before we went to sleep + * - descriptor tables + * - control registers + * - segment registers + * - flags + * + * Note that it is critical that this function is declared inline. + * It was separated out from restore_state to make that function + * a little clearer, but it needs to be inlined because we won't have a + * stack when we get here (so we can't push a return address). + */ +static inline void restore_processor_context(void) +{ + /* + * first restore %ds, so we can access our data properly + */ + asm volatile (".align 4"); + asm volatile ("movw %0, %%ds" :: "r" ((u16)__KERNEL_DS)); + + + /* + * control registers + */ + asm volatile ("movl %0, %%cr4" :: "r" (suspend2_saved_context.cr4)); + asm volatile ("movl %0, %%cr3" :: "r" (suspend2_saved_context.cr3)); + asm volatile ("movl %0, %%cr2" :: "r" (suspend2_saved_context.cr2)); + asm volatile ("movl %0, %%cr0" :: "r" (suspend2_saved_context.cr0)); + + /* + * segment registers + */ + asm volatile ("movw %0, %%es" :: "r" (suspend2_saved_context.es)); + asm volatile ("movw %0, %%fs" :: "r" (suspend2_saved_context.fs)); + asm volatile ("movw %0, %%gs" :: "r" (suspend2_saved_context.gs)); + asm volatile ("movw %0, %%ss" :: "r" (suspend2_saved_context.ss)); + + /* + * the other general registers + * + * note that even though gcc has constructs to specify memory + * input into certain registers, it will try to be too smart + * and save them at the beginning of the function. This is esp. + * bad since we don't have a stack set up when we enter, and we + * want to preserve the values on exit. So, we set them manually. + */ + asm volatile ("movl %0, %%esp" :: "m" (suspend2_saved_context.esp)); + asm volatile ("movl %0, %%ebp" :: "m" (suspend2_saved_context.ebp)); + asm volatile ("movl %0, %%eax" :: "m" (suspend2_saved_context.eax)); + asm volatile ("movl %0, %%ebx" :: "m" (suspend2_saved_context.ebx)); + asm volatile ("movl %0, %%ecx" :: "m" (suspend2_saved_context.ecx)); + asm volatile ("movl %0, %%edx" :: "m" (suspend2_saved_context.edx)); + asm volatile ("movl %0, %%esi" :: "m" (suspend2_saved_context.esi)); + asm volatile ("movl %0, %%edi" :: "m" (suspend2_saved_context.edi)); + + /* + * now restore the descriptor tables to their proper values + * ltr is done in fix_processor_context(). + */ + + asm volatile ("lgdt (%0)" :: "m" (suspend2_saved_context.gdt_limit)); + asm volatile ("lidt (%0)" :: "m" (suspend2_saved_context.idt_limit)); + asm volatile ("lldt (%0)" :: "m" (suspend2_saved_context.ldt)); + + fix_processor_context(); + + /* + * the flags + */ + asm volatile ("pushl %0 ; popfl" :: "m" (suspend2_saved_context.eflags)); + + do_fpu_end(); +} + +#if defined(CONFIG_SUSPEND2) || defined(CONFIG_SMP) +extern atomic_t suspend_cpu_counter __nosavedata; +extern unsigned char * my_saved_context __nosavedata; +static unsigned long c_loops_per_jiffy_ref[NR_CPUS] __nosavedata; +#endif + +#ifdef CONFIG_SUSPEND2 +#ifndef CONFIG_SMP +extern unsigned long loops_per_jiffy; +volatile static unsigned long cpu_khz_ref __nosavedata = 0; +#endif + +/* + * APIC support: These routines save the APIC + * configuration for the CPU on which they are + * being executed + */ +extern void suspend_apic_save_state(void); +extern void suspend_apic_reload_state(void); + +#ifdef CONFIG_SMP +/* ------------------------------------------------ + * BEGIN Irq affinity code, based on code from LKCD. + * + * IRQ affinity support: + * Save and restore IRQ affinities, and set them + * all to CPU 0. + * + * Section between dashes taken from LKCD code. + * Perhaps we should be working toward a shared library + * of such routines for kexec, lkcd, software suspend + * and whatever other similar projects there are? + */ + +extern irq_desc_t irq_desc[]; +extern cpumask_t irq_affinity[]; +extern cpumask_t saved_affinity[NR_IRQS]; + +/* + * Routine to save the old irq affinities and change affinities of all irqs to + * the dumping cpu. + */ +static void save_and_set_irq_affinity(void) +{ + int i; + int cpu = _smp_processor_id(); + + memcpy(saved_affinity, irq_affinity, NR_IRQS * sizeof(cpumask_t)); + for (i = 0; i < NR_IRQS; i++) { + if (irq_desc[i].handler == NULL) + continue; + irq_affinity[i] = cpumask_of_cpu(cpu); + if (irq_desc[i].handler->set_affinity != NULL) + irq_desc[i].handler->set_affinity(i, irq_affinity[i]); + } +} + +/* + * Restore old irq affinities. + */ +static void reset_irq_affinity(void) +{ + int i; + + memcpy(irq_affinity, saved_affinity, NR_IRQS * sizeof(cpumask_t)); + for (i = 0; i < NR_IRQS; i++) { + if (irq_desc[i].handler == NULL) + continue; + if (irq_desc[i].handler->set_affinity != NULL) + irq_desc[i].handler->set_affinity(i, irq_affinity[i]); + } +} + +/* + * END of IRQ affinity code, based on LKCD code. + * ----------------------------------------------------------------- + */ +#else +#define save_and_set_irq_affinity() do { } while(0) +#define reset_irq_affinity() do { } while(0) +#endif + +static inline void suspend2_pre_copy(void) +{ + /* + * Save the irq affinities before we freeze the + * other processors! + */ + save_and_set_irq_affinity(); + mtrr_save(); +} + +static inline void suspend2_post_copy(void) +{ +} + +static inline void suspend2_pre_copyback(void) +{ + /* We want to run from swsusp_pg_dir, since swsusp_pg_dir is stored in constant + * place in memory + */ + + __asm__( "movl %%ecx,%%cr3\n" ::"c"(__pa(swsusp_pg_dir))); + /* Send all IRQs to CPU 0. We will replace the saved affinities + * with the suspend-time ones when we copy the original kernel + * back in place + */ + save_and_set_irq_affinity(); + + c_loops_per_jiffy_ref[_smp_processor_id()] = current_cpu_data.loops_per_jiffy; +#ifndef CONFIG_SMP + cpu_khz_ref = cpu_khz; + c_loops_per_jiffy_ref[_smp_processor_id()] = loops_per_jiffy; +#endif + +} + +static inline void suspend2_restore_processor_context(void) +{ + restore_processor_context(); +} + +static inline void suspend2_flush_caches(void) +{ + cpu_clear(_smp_processor_id(), per_cpu(cpu_tlbstate, _smp_processor_id()).active_mm->cpu_vm_mask); + wbinvd(); + __flush_tlb_all(); + +} + +static inline void suspend2_post_copyback(void) +{ + mtrr_restore_one_cpu(); + + /* Get other CPUs to restore their contexts and flush their tlbs. */ + clear_suspend_state(SUSPEND_FREEZE_SMP); + + do { + cpu_relax(); + barrier(); + } while (atomic_read(&suspend_cpu_counter)); + mtrr_restore_finish(); + + BUG_ON(!irqs_disabled()); + + /* put the irq affinity tables back */ + reset_irq_affinity(); + + current_cpu_data.loops_per_jiffy = c_loops_per_jiffy_ref[_smp_processor_id()]; +#ifndef CONFIG_SMP + loops_per_jiffy = c_loops_per_jiffy_ref[_smp_processor_id()]; + cpu_khz = cpu_khz_ref; +#endif +} + +#endif diff -ruNp 500-version-specific-i386.patch-old/include/asm-i386/suspend.h 500-version-specific-i386.patch-new/include/asm-i386/suspend.h --- 500-version-specific-i386.patch-old/include/asm-i386/suspend.h 2005-06-20 11:47:15.000000000 +1000 +++ 500-version-specific-i386.patch-new/include/asm-i386/suspend.h 2005-07-04 23:14:19.000000000 +1000 @@ -3,6 +3,7 @@ * Based on code * Copyright 2001 Patrick Mochel <[EMAIL PROTECTED]> */ +#include <linux/errno.h> #include <asm/desc.h> #include <asm/i387.h>
- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/