cpr_wakecode.s revision 2df1fe9ca32bb227b9158c67f5c00b54c20b10fd
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License (the "License").
* You may not use this file except in compliance with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright 2007 Sun Microsystems, Inc. All rights reserved.
* Use is subject to license terms.
*/
#pragma ident "%Z%%M% %I% %E% SMI"
#include <sys/asm_linkage.h>
#include <sys/asm_misc.h>
#include <sys/regset.h>
#include <sys/privregs.h>
#include <sys/x86_archext.h>
#include <sys/cpr_wakecode.h>
#if !defined(__lint)
#include <sys/segments.h>
#include "assym.h"
#endif
/*
* This file contains the low level routines involved in getting
* into and out of ACPI S3, including those needed for restarting
* the non-boot cpus.
*
* Our assumptions:
*
* Our actions:
*
*/
#if defined(lint) || defined(__lint)
/*ARGSUSED*/
int
wc_save_context(wc_cpu_t *pcpu)
{ return 0; }
#else /* lint */
#if defined(__GNU_AS__)
NOTHING AT ALL YET!
#else /* !defined(__GNU_AS__) */
#if defined(__amd64)
ENTRY_NP(wc_save_context)
movq (%rsp), %rdx / return address
movq %rdx, WC_RETADDR(%rdi)
/*
* C calling convention with no local variables, just use 1st arg ie %rdi
* and the registers. Therefore push caller's fp, set out fp to be sp and
* push %r12, %r13 %r14. At function end unwind this by: popping %r14, %r13
* %r14, restore the sp from fp and pop caller's fp.
*/
pushq %rbp
movq %rsp,%rbp
pushq %r12
pushq %r13
pushq %r14
movq %rdi, WC_VIRTADDR(%rdi)
movq %rdi, WC_RDI(%rdi)
movq %rdx, WC_RDX(%rdi)
/ stash everything else we need
sgdt WC_GDT(%rdi)
sidt WC_IDT(%rdi)
sldt WC_LDT(%rdi)
str WC_TR(%rdi)
movq %cr0, %rdx
movq %rdx, WC_CR0(%rdi)
movq %cr3, %rdx
movq %rdx, WC_CR3(%rdi)
movq %cr4, %rdx
movq %rdx, WC_CR4(%rdi)
movq %cr8, %rdx
movq %rdx, WC_CR8(%rdi)
movq %r8, WC_R8(%rdi)
movq %r9, WC_R9(%rdi)
movq %r10, WC_R10(%rdi)
movq %r11, WC_R11(%rdi)
movq %r12, WC_R12(%rdi)
movq %r13, WC_R13(%rdi)
movq %r14, WC_R14(%rdi)
movq %r15, WC_R15(%rdi)
movq %rax, WC_RAX(%rdi)
movq %rbp, WC_RBP(%rdi)
movq %rbx, WC_RBX(%rdi)
movq %rcx, WC_RCX(%rdi)
movq %rsi, WC_RSI(%rdi)
movq %rsp, WC_RSP(%rdi)
movw %ss, WC_SS(%rdi)
movw %cs, WC_CS(%rdi)
movw %ds, WC_DS(%rdi)
movw %es, WC_ES(%rdi)
movq $0, %rcx / save %fs register
movw %fs, %cx
movq %rcx, WC_FS(%rdi)
movl $MSR_AMD_FSBASE, %ecx
rdmsr
movl %eax, WC_FSBASE(%rdi)
movl %edx, WC_FSBASE+4(%rdi)
movq $0, %rcx / save %gs register
movw %gs, %cx
movq %rcx, WC_GS(%rdi)
movl $MSR_AMD_GSBASE, %ecx / save gsbase msr
rdmsr
movl %eax, WC_GSBASE(%rdi)
movl %edx, WC_GSBASE+4(%rdi)
movl $MSR_AMD_KGSBASE, %ecx / save kgsbase msr
rdmsr
movl %eax, WC_KGSBASE(%rdi)
movl %edx, WC_KGSBASE+4(%rdi)
pushfq
popq WC_EFLAGS(%rdi)
/*
* call save_stack(cpup)
* NB %rdi is the first arguemnt to both wc_save_context() and save_stack()
* so it must not be modified during either of these calls.
* The pushq will decrement the value of %rsp
* we need to push the %rbp because it is the frame pointer and we need
* to use the C calling convention
*/
pushq %rbp
call *save_stack_func
popq %rbp
wbinvd / flush the cache
movq $1, %rax / at suspend return 1
/ see comment at function enter
popq %r14
popq %r13
popq %r12
leave
ret
SET_SIZE(wc_save_context)
#elif defined(__i386)
ENTRY_NP(wc_save_context)
movl 4(%esp), %eax / wc_cpu_t *
movl %eax, WC_VIRTADDR(%eax)
movl (%esp), %edx / return address
movl %edx, WC_RETADDR(%eax)
str WC_TR(%eax) / stash everything else we need
sgdt WC_GDT(%eax)
sldt WC_LDT(%eax)
sidt WC_IDT(%eax)
movl %cr0, %edx
movl %edx, WC_CR0(%eax)
movl %cr3, %edx
movl %edx, WC_CR3(%eax)
movl %cr4, %edx
movl %edx, WC_CR4(%eax)
movl %ebx, WC_EBX(%eax)
movl %edi, WC_EDI(%eax)
movl %esi, WC_ESI(%eax)
movl %ebp, WC_EBP(%eax)
movl %esp, WC_ESP(%eax)
movw %ss, WC_SS(%eax)
movw %cs, WC_CS(%eax)
movw %ds, WC_DS(%eax)
movw %es, WC_ES(%eax)
movw %fs, WC_FS(%eax)
movw %gs, WC_GS(%eax)
pushfl
popl WC_EFLAGS(%eax)
wbinvd / flush the cache
movl $1, %eax / at suspend return 1
ret
SET_SIZE(wc_save_context)
#endif /* __amd64 */
#endif /* __GNU_AS__ */
#endif /* lint */
/*
* Our assumptions:
* - We are running in real mode.
* - Interrupts are disabled.
*
* Our actions:
* - We start using our GDT by loading correct values in the
* selector registers (cs=KCS_SEL, ds=es=ss=KDS_SEL, fs=KFS_SEL,
* gs=KGS_SEL).
* - We change over to using our IDT.
* - We load the default LDT into the hardware LDT register.
* - We load the default TSS into the hardware task register.
* - We restore registers
* - We return to original caller (a la setjmp)
*/
#if defined(lint) || defined(__lint)
void
wc_rm_start(void)
{}
void
wc_rm_end(void)
{}
#else /* lint */
#if defined(__GNU_AS__)
NOTHING AT ALL YET!
#else /* __GNU_AS__ */
#if defined(__amd64)
ENTRY_NP(wc_rm_start)
/*
* For vulcan as we need to do a .code32 and mentally invert the
* meaning of the addr16 and data16 prefixes to get 32-bit access when
* generating code to be executed in 16-bit mode (sigh...)
*/
.code32
cli
movw %cs, %ax
movw %ax, %ds / establish ds ...
movw %ax, %ss / ... and ss:esp
D16 movl $WC_STKSTART, %esp
/ using the following value blows up machines! - DO NOT USE
/ D16 movl 0xffc, %esp
#define LED 1
#define SERIAL 1
#if LED
D16 movl $0x80, %edx
D16 movb $0xd1, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x61, %al
outb (%dx)
#endif
D16 call cominit
/*
* Enable protected-mode, write protect, and alignment mask
* %cr0 has already been initialsed to zero
*/
movl %cr0, %eax
D16 orl $[CR0_PE|CR0_WP|CR0_AM], %eax
movl %eax, %cr0
/*
* Do a jmp immediately after writing to cr0 when enabling protected
* mode to clear the real mode prefetch queue (per Intel's docs)
*/
jmp pestart
pestart:
#if LED
D16 movl $0x80, %edx
D16 movb $0xd2, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x62, %al
outb (%dx)
#endif
/*
* 16-bit protected mode is now active, so prepare to turn on long
* mode
*/
#if LED
D16 movl $0x80, %edx
D16 movb $0xd3, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x63, %al
outb (%dx)
#endif
/*
* Add any initial cr4 bits
*/
movl %cr4, %eax
A16 D16 orl CR4OFF, %eax
/*
* Enable PAE mode (CR4.PAE)
*/
D16 orl $CR4_PAE, %eax
movl %eax, %cr4
#if LED
D16 movl $0x80, %edx
D16 movb $0xd4, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x64, %al
outb (%dx)
#endif
/*
* Point cr3 to the 64-bit long mode page tables.
*
* Note that these MUST exist in 32-bit space, as we don't have
* a way to load %cr3 with a 64-bit base address for the page tables
* until the CPU is actually executing in 64-bit long mode.
*/
A16 D16 movl CR3OFF, %eax
movl %eax, %cr3
/*
* Set long mode enable in EFER (EFER.LME = 1)
*/
D16 movl $MSR_AMD_EFER, %ecx
rdmsr
D16 orl $AMD_EFER_LME, %eax
wrmsr
#if LED
D16 movl $0x80, %edx
D16 movb $0xd5, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x65, %al
outb (%dx)
#endif
/*
* Finally, turn on paging (CR0.PG = 1) to activate long mode.
*/
movl %cr0, %eax
D16 orl $CR0_PG, %eax
movl %eax, %cr0
/*
* The instruction after enabling paging in CR0 MUST be a branch.
*/
jmp long_mode_active
long_mode_active:
#if LED
D16 movl $0x80, %edx
D16 movb $0xd6, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x66, %al
outb (%dx)
#endif
/*
* Long mode is now active but since we're still running with the
* original 16-bit CS we're actually in 16-bit compatability mode.
*
* We have to load an intermediate GDT and IDT here that we know are
* in 32-bit space before we can use the kernel's GDT and IDT, which
* may be in the 64-bit address space, and since we're in compatability
* mode, we only have access to 16 and 32-bit instructions at the
* moment.
*/
A16 D16 lgdt TEMPGDTOFF /* load temporary GDT */
A16 D16 lidt TEMPIDTOFF /* load temporary IDT */
/*
* Do a far transfer to 64-bit mode. Set the CS selector to a 64-bit
* long mode selector (CS.L=1) in the temporary 32-bit GDT and jump
* to the real mode platter address of wc_long_mode_64 as until the
* 64-bit CS is in place we don't have access to 64-bit instructions
* and thus can't reference a 64-bit %rip.
*/
#if LED
D16 movl $0x80, %edx
D16 movb $0xd7, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x67, %al
outb (%dx)
#endif
D16 pushl $TEMP_CS64_SEL
A16 D16 pushl LM64OFF
D16 lret
/*
* Support routine to re-initialize VGA subsystem
*/
vgainit:
D16 ret
/*
* Support routine to re-initialize keyboard (which is USB - help!)
*/
kbdinit:
D16 ret
/*
* Support routine to re-initialize COM ports to something sane
*/
cominit:
/ init COM1 & COM2
xorl %edx, %edx / select COM1
D16 movl $0xe3, %eax / ah=0; al=(9600bd|8_bit|nopar)
int $0x14
D16 movl $1, %edx / select COM2
D16 movl $0xe3, %eax / ah=0; al=(9600bd|8_bit|nopar)
int $0x14
D16 ret
.code64
/*
* Support routine to re-initialize COM ports to something sane
*/
cominit64:
/ init COM1 & COM2
xorq %rdx, %rdx / select COM1
movq $0xe3, %rax / ah=0; al=(9600bd|8_bit|nopar)
int $0x14
movq $1, %rdx / select COM2
movq $0xe3, %rax / ah=0; al=(9600bd|8_bit|nopar)
int $0x14
ret
.globl wc_long_mode_64
wc_long_mode_64:
#if LED
movw $0x80, %dx
movb $0xd8, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movb $0x68, %al
outb (%dx)
#endif
/*
* We are now running in long mode with a 64-bit CS (EFER.LMA=1,
* CS.L=1) so we now have access to 64-bit instructions.
*
* First, set the 64-bit GDT base.
*/
.globl rm_platter_pa
movl rm_platter_pa, %eax
lgdtq GDTROFF(%rax) /* load 64-bit GDT */
/*
* Save the CPU number in %r11; get the value here since it's saved in
* the real mode platter.
*/
/ JAN
/ the following is wrong! need to figure out MP systems
/ movl CPUNOFF(%rax), %r11d
/*
* Add rm_platter_pa to %rsp to point it to the same location as seen
* from 64-bit mode.
*/
addq %rax, %rsp
/*
* Now do an lretq to load CS with the appropriate selector for the
* kernel's 64-bit GDT and to start executing 64-bit setup code at the
* virtual address where boot originally loaded this code rather than
* the copy in the real mode platter's rm_code array as we've been
* doing so far.
*/
#if LED
movw $0x80, %dx
movb $0xd9, %al
outb (%dx)
#endif
/ JAN this should produce 'i' but we get 'g' instead ???
#if SERIAL
movw $0x3f8, %dx
movb $0x69, %al
outb (%dx)
#endif
pushq $KCS_SEL
pushq $kernel_wc_code
lretq
.globl kernel_wc_code
kernel_wc_code:
#if LED
movw $0x80, %dx
movb $0xda, %al
outb (%dx)
#endif
/ JAN this should produce 'j' but we get 'g' instead ???
#if SERIAL
movw $0x3f8, %dx
movb $0x6a, %al
outb (%dx)
#endif
/*
* Complete the balance of the setup we need to before executing
* 64-bit kernel code (namely init rsp, TSS, LGDT, FS and GS).
*/
.globl rm_platter_va
movq rm_platter_va, %rbx
addq $WC_CPU, %rbx
#if LED
movw $0x80, %dx
movb $0xdb, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movw $0x6b, %ax
outb (%dx)
#endif
/*
* restore the rest of the registers
*/
lidtq WC_IDT(%rbx)
#if LED
movw $0x80, %dx
movb $0xdc, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movw $0x6c, %ax
outb (%dx)
#endif
/*
* restore the rest of the registers
*/
movw $KDS_SEL, %ax
movw %ax, %ds
movw %ax, %es
movw %ax, %ss
/*
* Before proceeding, enable usage of the page table NX bit if
* that's how the page tables are set up.
*/
movl x86_feature, %ecx
andl $X86_NX, %ecx
jz 1f
movl $MSR_AMD_EFER, %ecx
rdmsr
orl $AMD_EFER_NXE, %eax
wrmsr
1:
movq WC_CR4(%rbx), %rax / restore full cr4 (with Global Enable)
movq %rax, %cr4
lldt WC_LDT(%rbx)
movzwq WC_TR(%rbx), %rax / clear TSS busy bit
addq WC_GDT+2(%rbx), %rax
andl $0xfffffdff, 4(%rax)
movq 4(%rax), %rcx
ltr WC_TR(%rbx)
#if LED
movw $0x80, %dx
movb $0xdd, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movw $0x6d, %ax
outb (%dx)
#endif
/ restore %fsbase %gsbase %kgbase registers using wrmsr instruction
movq WC_FS(%rbx), %rcx / restore fs register
movw %cx, %fs
movl $MSR_AMD_FSBASE, %ecx
movl WC_FSBASE(%rbx), %eax
movl WC_FSBASE+4(%rbx), %edx
wrmsr
movq WC_GS(%rbx), %rcx / restore gs register
movw %cx, %gs
movl $MSR_AMD_GSBASE, %ecx / restore gsbase msr
movl WC_GSBASE(%rbx), %eax
movl WC_GSBASE+4(%rbx), %edx
wrmsr
movl $MSR_AMD_KGSBASE, %ecx / restore kgsbase msr
movl WC_KGSBASE(%rbx), %eax
movl WC_KGSBASE+4(%rbx), %edx
wrmsr
movq WC_CR0(%rbx), %rdx
movq %rdx, %cr0
movq WC_CR3(%rbx), %rdx
movq %rdx, %cr3
movq WC_CR8(%rbx), %rdx
movq %rdx, %cr8
#if LED
movw $0x80, %dx
movb $0xde, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movb $0x6e, %al
outb (%dx)
#endif
/ dummy up a stck so we can make C function calls
movq WC_RSP(%rbx), %rsp
/*
* APIC initialization (we dummy up a stack so we can make this call)
*/
pushq $0 /* null frame pointer terminates stack trace */
movq %rsp, %rbp /* stack aligned on 16-byte boundary */
call *ap_mlsetup
call *cpr_start_cpu_func
/ restore %rbx to the value it ahd before we called the functions above
movq rm_platter_va, %rbx
addq $WC_CPU, %rbx
movq WC_R8(%rbx), %r8
movq WC_R9(%rbx), %r9
movq WC_R10(%rbx), %r10
movq WC_R11(%rbx), %r11
movq WC_R12(%rbx), %r12
movq WC_R13(%rbx), %r13
movq WC_R14(%rbx), %r14
movq WC_R15(%rbx), %r15
/ movq WC_RAX(%rbx), %rax
movq WC_RBP(%rbx), %rbp
movq WC_RCX(%rbx), %rcx
/ movq WC_RDX(%rbx), %rdx
movq WC_RDI(%rbx), %rdi
movq WC_RSI(%rbx), %rsi
/ assume that %cs does not need to be restored
/ %ds, %es & %ss are ignored in 64bit mode
movw WC_SS(%rbx), %ss
movw WC_DS(%rbx), %ds
movw WC_ES(%rbx), %es
#if LED
movw $0x80, %dx
movb $0xdf, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movb $0x6f, %al
outb (%dx)
#endif
movq WC_RBP(%rbx), %rbp
movq WC_RSP(%rbx), %rsp
#if LED
movw $0x80, %dx
movb $0xe0, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movb $0x70, %al
outb (%dx)
#endif
movq WC_RCX(%rbx), %rcx
pushq WC_EFLAGS(%rbx) / restore flags
popfq
#if LED
movw $0x80, %dx
movb $0xe1, %al
outb (%dx)
#endif
#if SERIAL
movw $0x3f8, %dx
movb $0x71, %al
outb (%dx)
#endif
/*
* can not use outb after this point, because doing so would mean using
* %dx which would modify %rdx which is restored here
*/
movq %rbx, %rax
movq WC_RDX(%rax), %rdx
movq WC_RBX(%rax), %rbx
popq %r14
popq %r13
popq %r12
leave
movq WC_RETADDR(%rax), %rax
movq %rax, (%rsp) / return to caller of wc_save_context
xorl %eax, %eax / at wakeup return 0
ret
SET_SIZE(wc_rm_start)
ENTRY_NP(asmspin)
movl %edi, %ecx
A1:
loop A1
SET_SIZE(asmspin)
.globl wc_rm_end
wc_rm_end:
nop
#elif defined(__i386)
ENTRY_NP(wc_rm_start)
/entry: jmp entry / stop here for HDT
cli
movw %cs, %ax
movw %ax, %ds / establish ds ...
movw %ax, %ss / ... and ss:esp
D16 movl $WC_STKSTART, %esp
#define LED 1
#define SERIAL 1
#if LED
D16 movl $0x80, %edx
D16 movb $0xd1, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x61, %al
outb (%dx)
#endif
D16 call vgainit
D16 call kbdinit
D16 call cominit
#if LED
D16 movl $0x80, %edx
D16 movb $0xd2, %al
outb (%dx)
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x62, %al
outb (%dx)
#endif
D16 A16 movl $WC_CPU, %ebx / base add of wc_cpu_t
#if LED
D16 movb $0xd3, %al
outb $0x80
#endif
#if SERIAL
D16 movl $0x3f8, %edx
D16 movb $0x63, %al
outb (%dx)
#endif
D16 A16 movl %cs:WC_DS(%ebx), %edx / %ds post prot/paging transit
D16 movb $0xd4, %al
outb $0x80
D16 A16 lgdt %cs:WC_GDT(%ebx) / restore gdt and idtr
D16 A16 lidt %cs:WC_IDT(%ebx)
D16 movb $0xd5, %al
outb $0x80
D16 A16 movl %cs:WC_CR4(%ebx), %eax / restore cr4
D16 andl $-1!CR4_PGE, %eax / don't set Global Enable yet
movl %eax, %cr4
D16 movb $0xd6, %al
outb $0x80
D16 A16 movl %cs:WC_CR3(%ebx), %eax / set PDPT
movl %eax, %cr3
D16 movb $0xd7, %al
outb $0x80
D16 A16 movl %cs:WC_CR0(%ebx), %eax / enable prot/paging, etc.
movl %eax, %cr0
D16 movb $0xd8, %al
outb $0x80
D16 A16 movl %cs:WC_VIRTADDR(%ebx), %ebx / virtaddr of wc_cpu_t
D16 movb $0xd9, %al
outb $0x80
D16 movb $0xda, %al
outb $0x80
jmp flush / flush prefetch queue
flush:
D16 pushl $KCS_SEL
D16 pushl $kernel_wc_code
D16 lret / re-appear at kernel_wc_code
/*
* Support routine to re-initialize VGA subsystem
*/
vgainit:
D16 ret
/*
* Support routine to re-initialize keyboard (which is USB - help!)
*/
kbdinit:
D16 ret
/*
* Support routine to re-initialize COM ports to something sane for debug output
*/
cominit:
/ init COM1 & COM2
xorl %edx, %edx / select COM1
D16 movl $0xe3, %eax / ah=0; al=(9600bd|8_bit|nopar)
int $0x14
D16 movl $1, %edx / select COM2
D16 movl $0xe3, %eax / ah=0; al=(9600bd|8_bit|nopar)
int $0x14
D16 ret
.globl wc_rm_end
wc_rm_end:
nop
.globl kernel_wc_code
kernel_wc_code:
/ At this point we are with kernel's cs and proper eip.
/ We will be executing not from the copy in real mode platter,
/ but from the original code where boot loaded us.
/ By this time GDT and IDT are loaded as is cr0, cr3 and cr4.
/ %ebx is wc_cpu
/ %dx is our ds
D16 movb $0xdb, %al
outb $0x80
/ got here OK
movw %dx, %ds / $KDS_SEL
movb $0xdc, %al
outb $0x80
movl $MSR_AMD_EFER, %ecx / re-enable NX bit
rdmsr
orl $AMD_EFER_NXE, %eax
wrmsr
movl WC_CR4(%ebx), %eax / restore full cr4 (with Global Enable)
movl %eax, %cr4
lldt WC_LDT(%ebx) / $LDT_SEL
movzwl WC_TR(%ebx), %eax / clear TSS busy bit
addl WC_GDT+2(%ebx), %eax
andl $-1!0x200, 4(%eax)
ltr WC_TR(%ebx) / $UTSS_SEL
movw WC_SS(%ebx), %ss / lssl WC_ESP(%ebx), %esp
movl WC_ESP(%ebx), %esp / ^ don't use, asm busted!
movl WC_RETADDR(%ebx), %eax / return to caller of wc_save_context
movl %eax, (%esp)
movw WC_ES(%ebx), %es / restore segment registers
movw WC_FS(%ebx), %fs
movw WC_GS(%ebx), %gs
/*
* APIC initialization
*/
call *ap_mlsetup
call *cpr_start_cpu_func
pushl WC_EFLAGS(%ebx) / restore flags
popfl
movl WC_EDI(%ebx), %edi / restore general registers
movl WC_ESI(%ebx), %esi
movl WC_EBP(%ebx), %ebp
movl WC_EBX(%ebx), %ebx
/exit: jmp exit / stop here for HDT
xorl %eax, %eax / at wakeup return 0
ret
SET_SIZE(wc_rm_start)
#endif /* defined(__amd64) */
#endif /* !defined(__GNU_AS__) */
#endif /* lint */