/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibSse2/X64/ |
H A D | SetMem64.asm | 36 xchg rcx, rdx ; rcx <- Count & rdx <- Buffer 40 mov [rdx], r8 41 add rdx, 8 48 movntdq [rdx], xmm0 49 lea rdx, [rdx + 16] 54 mov [rdx], r8
|
H A D | SetMem64.S | 39 xchgq %rdx, %rcx # rcx <- Count & rdx <- Buffer 43 movq %r8, (%rdx) 44 addq $8, %rdx 51 movntdq %xmm0, (%rdx) 52 leaq 16(%rdx), %rdx 57 movq %r8, (%rdx)
|
H A D | SetMem32.asm | 43 cmp rcx, rdx 44 cmova rcx, rdx 45 sub rdx, rcx 48 mov rcx, rdx
|
H A D | ZeroMem.asm | 41 cmp rcx, rdx 42 cmova rcx, rdx 43 sub rdx, rcx 46 mov rcx, rdx
|
H A D | SetMem32.S | 48 cmpq %rdx, %rcx 49 cmova %rdx, %rcx 50 subq %rcx, %rdx 53 movq %rdx, %rcx
|
H A D | ZeroMem.S | 45 cmpq %rdx, %rcx 46 cmova %rdx, %rcx 47 subq %rcx, %rdx 50 movq %rdx, %rcx
|
H A D | SetMem.asm | 42 cmp rcx, rdx 43 cmova rcx, rdx 44 sub rdx, rcx 47 mov rcx, rdx 48 and rdx, 15
|
/vbox/src/libs/zlib-1.2.6/contrib/amd64/ |
H A D | amd64-match.S | 269 mov $(-MAX_MATCH_8), %rdx 275 prefetcht1 (%windowbestlen, %rdx) 276 prefetcht1 (%prev, %rdx) 280 * adjust %rdx so that it is offset to the exact byte that mismatched. 293 movdqu (%windowbestlen, %rdx), %xmm1 294 movdqu (%prev, %rdx), %xmm2 296 movdqu 16(%windowbestlen, %rdx), %xmm3 297 movdqu 16(%prev, %rdx), %xmm4 299 movdqu 32(%windowbestlen, %rdx), %xmm5 300 movdqu 32(%prev, %rdx), [all...] |
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibMmx/X64/ |
H A D | SetMem32.asm | 37 xchg rcx, rdx ; rcx <- Count rdx <- Buffer 42 DB 0fh, 0e7h, 02h ; movntq [rdx], mm0 43 lea rdx, [rdx + 8] ; use "lea" to avoid flag changes 48 DB 0fh, 7eh, 02h ; movd [rdx], mm0
|
H A D | SetMem32.S | 41 xchgq %rdx, %rcx # rcx <- Count rdx <- Buffer 46 movntq %mm0, (%rdx) 47 leaq 8(%rdx), %rdx # use "lea" to avoid flag changes 52 movd %mm0, (%rdx)
|
H A D | SetMem64.S | 40 xchg %rdx, %rcx #rcx <- Count 42 movntq %mm0, (%rdx) 43 addq $8, %rdx
|
H A D | CompareMem.asm | 45 mov rdi, rdx 49 movzx rdx, byte ptr [rdi - 1] 50 sub rax, rdx
|
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibOptDxe/X64/ |
H A D | ZeroMem.asm | 37 mov rcx, rdx ; rcx = Count 39 and rdx, 7 ; rdx = rdx & 7 42 mov rcx, rdx ; rcx = rdx
|
H A D | ZeroMem.S | 41 movq %rdx, %rcx 43 andq $7, %rdx 46 movq %rdx, %rcx
|
/vbox/src/VBox/Devices/PC/ipxe/src/arch/x86_64/core/linux/ |
H A D | linux_syscall.S | 15 movq %rdx, %rsi // C arg3 -> syscall arg2 16 movq %rcx, %rdx // C arg4 -> syscall arg3
|
/vbox/src/VBox/VMM/VMMR0/ |
H A D | CPUMR0UnusedA.asm | 82 fxrstor [rdx + CPUMCTX.fpu] 125 fxsave [rdx + CPUMCTX.fpu] 191 movdqa xmm0, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*0] 192 movdqa xmm1, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*1] 193 movdqa xmm2, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*2] 194 movdqa xmm3, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*3] 195 movdqa xmm4, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*4] 196 movdqa xmm5, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*5] 197 movdqa xmm6, [rdx + CPUMCTX.fpu + X86FXSTATE.aXMM + 16*6] 198 movdqa xmm7, [rdx [all...] |
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseLib/X64/ |
H A D | SwitchStack.asm | 30 ; (rdx) Context1 - Parameter1 for entry point. 41 mov rcx, rdx 42 mov rdx, r8 44 ; Reserve space for register parameters (rcx, rdx, r8 & r9) on the stack,
|
H A D | SwitchStack.S | 29 # (rdx) Context1 - Parameter1 for entry point. 44 mov %rdx, %rcx 45 mov %r8, %rdx 47 # Reserve space for register parameters (rcx, rdx, r8 & r9) on the stack,
|
/vbox/src/libs/zlib-1.2.6/contrib/gcc_gvmat64/ |
H A D | gvmat64.S | 68 ; register used : rax,rbx,rcx,rdx,rsi,rdi,r8,r9,r10,r11,r12 170 ; parameter 1 in rcx(deflate state s), param 2 in rdx (cur match) 176 ; rax, rcx, rdx, r8, r9, r10, and r11, which are scratch. 195 ; this clear high 32 bits of r8, which can be garbage in both r8 and rdx 419 mov rdx, 0xfffffffffffffef8 //; -(MAX_MATCH_8) 423 prefetcht1 [rsi+rdx] 424 prefetcht1 [rdi+rdx] 428 ;;; adjust rdx so that it is offset to the exact byte that mismatched. 443 mov rax, [rsi + rdx] 444 xor rax, [rdi + rdx] [all...] |
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibOptPei/X64/ |
H A D | SetMem.S | 42 xchgq %rdx, %rcx # rcx = Count, rdx = Buffer 44 movq %rdx, %rax # rax = Buffer
|
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibRepStr/X64/ |
H A D | SetMem.S | 42 xchgq %rdx, %rcx # rcx = Count, rdx = Buffer 44 movq %rdx, %rax # rax = Buffer
|
/vbox/src/VBox/Runtime/common/asm/ |
H A D | ASMXRstor.asm | 40 ; msc=rdx, gcc=rsi, x86=[esp+8] 45 mov rdx, rdx 46 shr rdx, 32 50 mov rdx, rsi 51 shr rdx, 32
|
H A D | ASMXSave.asm | 40 ; msc=rdx, gcc=rsi, x86=[esp+8] 45 mov rdx, rdx 46 shr rdx, 32 50 mov rdx, rsi 51 shr rdx, 32
|
/vbox/src/recompiler/ |
H A D | VBoxREMWrapperA.asm | 44 push rdx 72 pop rdx 114 mov rdx, rax 187 mov rdx, rsi 208 mov r8, rdx 209 mov rdx, rsi 231 mov r8, rdx 232 mov rdx, rsi 255 mov r8, rdx 256 mov rdx, rs [all...] |
/vbox/src/VBox/Devices/EFI/Firmware/MdeModulePkg/Universal/EbcDxe/X64/ |
H A D | EbcLowLevel.S | 48 sub %rdx, %r8 58 mov 0x8(%rsp), %rdx 75 mov %rdx, 0x10(%rsp) 97 mov 0x8(%rsp), %rdx 112 mov %rdx, %r8 113 mov %rcx, %rdx
|