/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibMmx/X64/ |
H A D | CopyMem.S | 39 pushq %rsi 41 movq %rdx, %rsi # rsi <- Source 43 leaq -1(%rsi, %r8,), %r9 # r9 <- End of Source 44 cmpq %rdi, %rsi 56 movq (%rsi), %mm0 58 addq $8, %rsi 65 movq %r9, %rsi # rsi <- End of Source 73 popq %rsi [all...] |
H A D | CopyMem.asm | 35 InternalMemCopyMem PROC USES rsi rdi 36 mov rsi, rdx ; rsi <- Source 38 lea r9, [rsi + r8 - 1] ; r9 <- End of Source 39 cmp rsi, rdi 51 DB 0fh, 6fh, 06h ; movd mm0, [rsi] 53 add rsi, 8 60 mov rsi, r9 ; rsi <- End of Source
|
H A D | CompareMem.S | 47 pushq %rsi 49 movq %rcx, %rsi 53 movzbq -1(%rsi), %rax 57 popq %rsi
|
H A D | CompareMem.asm | 43 InternalMemCompareMem PROC USES rsi rdi 44 mov rsi, rcx 48 movzx rax, byte ptr [rsi - 1]
|
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibOptDxe/X64/ |
H A D | CopyMem.S | 38 pushq %rsi 40 movq %rdx, %rsi # rsi <- Source 42 leaq -1(%rsi,%r8,), %r9 # r9 <- Last byte of Source 43 cmpq %rdi, %rsi 51 andq $15, %rcx # rcx + rsi should be 16 bytes aligned 64 movdqu (%rsi), %xmm0 # rsi may not be 16-byte aligned 66 addq $16, %rsi 73 movq %r9, %rsi # rs [all...] |
H A D | CopyMem.asm | 35 InternalMemCopyMem PROC USES rsi rdi 36 mov rsi, rdx ; rsi <- Source 38 lea r9, [rsi + r8 - 1] ; r9 <- Last byte of Source 39 cmp rsi, rdi 47 and rcx, 15 ; rcx + rsi should be 16 bytes aligned 60 movdqu xmm0, [rsi] ; rsi may not be 16-byte aligned 62 add rsi, 16 69 mov rsi, r [all...] |
H A D | CompareMem.S | 47 pushq %rsi 49 movq %rcx, %rsi 53 movzbq -1(%rsi) , %rax 57 popq %rsi
|
H A D | CompareMem.asm | 43 InternalMemCompareMem PROC USES rsi rdi 44 mov rsi, rcx 48 movzx rax, byte ptr [rsi - 1]
|
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibSse2/X64/ |
H A D | CopyMem.S | 39 pushq %rsi 41 movq %rdx, %rsi # rsi <- Source 43 leaq -1(%rsi, %r8,), %r9 # r9 <- Last byte of Source 44 cmpq %rdi, %rsi 52 andq $15, %rcx # rcx + rsi should be 16 bytes aligned 65 movdqu (%rsi), %xmm0 # rsi may not be 16-byte aligned 67 addq $16, %rsi 74 movq %r9, %rsi # rs [all...] |
H A D | CopyMem.asm | 35 InternalMemCopyMem PROC USES rsi rdi 36 mov rsi, rdx ; rsi <- Source 38 lea r9, [rsi + r8 - 1] ; r9 <- Last byte of Source 39 cmp rsi, rdi 47 and rcx, 15 ; rcx + rsi should be 16 bytes aligned 60 movdqu xmm0, [rsi] ; rsi may not be 16-byte aligned 62 add rsi, 16 69 mov rsi, r [all...] |
H A D | CompareMem.S | 47 pushq %rsi 49 movq %rcx, %rsi 53 movzbq -1(%rsi) , %rax 57 popq %rsi
|
H A D | CompareMem.asm | 43 InternalMemCompareMem PROC USES rsi rdi 44 mov rsi, rcx 48 movzx rax, byte ptr [rsi - 1]
|
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibOptPei/X64/ |
H A D | CopyMem.S | 39 pushq %rsi 41 movq %rdx, %rsi # rsi <- Source 43 leaq -1(%rsi, %r8,), %r9 # r9 <- End of Source 44 cmpq %rdi, %rsi 56 movq %r9, %rsi # rsi <- End of Source 64 popq %rsi
|
H A D | CopyMem.asm | 35 InternalMemCopyMem PROC USES rsi rdi 36 mov rsi, rdx ; rsi <- Source 38 lea r9, [rsi + r8 - 1] ; r9 <- End of Source 39 cmp rsi, rdi 51 mov rsi, r9 ; rsi <- End of Source
|
H A D | CompareMem.S | 47 pushq %rsi 49 movq %rcx, %rsi 53 movzbq -1(%rsi) , %rax 57 popq %rsi
|
H A D | CompareMem.asm | 43 InternalMemCompareMem PROC USES rsi rdi 44 mov rsi, rcx 48 movzx rax, byte ptr [rsi - 1]
|
/vbox/src/VBox/Devices/EFI/Firmware/MdePkg/Library/BaseMemoryLibRepStr/X64/ |
H A D | CopyMem.S | 39 pushq %rsi 41 movq %rdx, %rsi # rsi <- Source 43 leaq -1(%rsi, %r8,), %r9 # r9 <- End of Source 44 cmpq %rdi, %rsi 56 movq %r9, %rsi # rsi <- End of Source 64 popq %rsi
|
H A D | CopyMem.asm | 35 InternalMemCopyMem PROC USES rsi rdi 36 mov rsi, rdx ; rsi <- Source 38 lea r9, [rsi + r8 - 1] ; r9 <- End of Source 39 cmp rsi, rdi 51 mov rsi, r9 ; rsi <- End of Source
|
H A D | CompareMem.S | 47 pushq %rsi 49 movq %rcx, %rsi 53 movzbq -1(%rsi) , %rax 57 popq %rsi
|
H A D | CompareMem.asm | 43 InternalMemCompareMem PROC USES rsi rdi 44 mov rsi, rcx 48 movzx rax, byte ptr [rsi - 1]
|
/vbox/src/VBox/Devices/PC/ipxe/src/arch/x86_64/core/linux/ |
H A D | linux_syscall.S | 14 movq %rsi, %rdi // C arg2 -> syscall arg1 15 movq %rdx, %rsi // C arg3 -> syscall arg2
|
H A D | linuxprefix.S | 12 movq %rsp, %rsi // argv -> C arg2
|
/vbox/src/libs/zlib-1.2.6/contrib/masmx64/ |
H A D | inffasx64.asm | 32 mov [rsp-8],rsi 47 mov rsi, [rsp+16] ; /* rsi = in */ 64 cmp r9, rsi 85 cmp r9, rsi 190 mov r8, rsi ; /* save in so from can use it's reg */ 198 mov rsi, rdi 199 sub rsi, r15 ; /* from = out - dist */ 205 mov al, [rsi] 209 mov rsi, r [all...] |
/vbox/src/libs/zlib-1.2.6/contrib/gcc_gvmat64/ |
H A D | gvmat64.S | 68 ; register used : rax,rbx,rcx,rdx,rsi,rdi,r8,r9,r10,r11,r12 181 ; param 1 in rdi, param 2 in rsi 194 ; mac: param 1 in rdi, param 2 rsi 283 lea rsi,[r10+r11] 298 cmp bx,word ptr [rsi + r8 - 1] 318 cmp bx,word ptr [rsi + r8 - 1] 337 cmp bx,word ptr [rsi + r8 - 1] 357 cmp bx,word ptr [rsi + r8 - 1] 399 cmp bx,word ptr [rsi + r8 - 1] 418 lea rsi,[r [all...] |
/vbox/src/recompiler/ |
H A D | VBoxREMWrapperA.asm | 45 push rsi 71 pop rsi 117 mov rsi, eax 187 mov rdx, rsi 209 mov rdx, rsi 232 mov rdx, rsi 256 mov rdx, rsi 281 mov rdx, rsi 308 mov rdx, rsi 337 mov rdx, rsi [all...] |