Home
last modified time | relevance | path

Searched refs:rsi (Results 1 – 25 of 138) sorted by relevance

123456

/linux/arch/x86/virt/vmx/tdx/
H A Dtdxcall.S52 movq TDX_MODULE_rcx(%rsi), %rcx
53 movq TDX_MODULE_rdx(%rsi), %rdx
54 movq TDX_MODULE_r8(%rsi), %r8
55 movq TDX_MODULE_r9(%rsi), %r9
56 movq TDX_MODULE_r10(%rsi), %r10
57 movq TDX_MODULE_r11(%rsi), %r11
72 movq TDX_MODULE_r12(%rsi), %r12
73 movq TDX_MODULE_r13(%rsi), %r13
80 pushq %rsi
85 movq TDX_MODULE_rsi(%rsi), %rsi
[all …]
/linux/arch/x86/lib/
H A Dmemmove_64.S34 cmp %rdi, %rsi
36 mov %rsi, %r8
69 movq 2*8(%rsi), %r9
70 movq 3*8(%rsi), %r8
71 leaq 4*8(%rsi), %rsi
101 movq (%rsi), %r11
103 leaq -8(%rsi, %rdx), %rsi
127 addq %rdx, %rsi
139 leaq -4*8(%rsi), %rsi
151 subq %rdx, %rsi
[all …]
H A Dcopy_page_64.S33 movq 0x8*0(%rsi), %rax
34 movq 0x8*1(%rsi), %rbx
35 movq 0x8*2(%rsi), %rdx
36 movq 0x8*3(%rsi), %r8
37 movq 0x8*4(%rsi), %r9
38 movq 0x8*5(%rsi), %r10
42 prefetcht0 5*64(%rsi)
53 leaq 64 (%rsi), %rsi
66 movq 0x8*3(%rsi), %r8
67 movq 0x8*4(%rsi), %r9
[all …]
H A Dmemcpy_64.S65 movq 0*8(%rsi), %r8
66 movq 1*8(%rsi), %r9
67 movq 2*8(%rsi), %r10
68 movq 3*8(%rsi), %r11
69 leaq 4*8(%rsi), %rsi
84 addq %rdx, %rsi
98 leaq -4*8(%rsi), %rsi
110 subq %rdx, %rsi
119 movq 0*8(%rsi), %r8
120 movq 1*8(%rsi), %r9
[all …]
H A Dcsum-copy_64.S107 movq %rbx, (%rsi)
109 movq %r8, 8(%rsi)
125 leaq 64(%rsi), %rsi
147 leaq 8(%rsi), %rsi /* preserve carry */
176 movw %bx, (%rsi)
178 leaq 2(%rsi), %rsi
190 movb %bl, (%rsi)
217 movw %bx, (%rsi)
220 leaq 2(%rsi), %rsi
230 leaq 4(%rsi), %rsi
[all …]
H A Dcopy_user_uncached_64.S40 10: movq (%rsi),%r8
41 11: movq 8(%rsi),%r9
48 30: movq 32(%rsi),%r8
57 addq $64,%rsi
104 addq $8,%rsi
123 addq $4,%rsi
132 addq $2,%rsi
172 inc %rsi
182 addq $2,%rsi
192 addq $4,%rsi
[all …]
H A Dcmpxchg16b_emu.S26 cmpq __percpu (%rsi), %rax
28 cmpq __percpu 8(%rsi), %rdx
32 movq %rbx, __percpu (%rsi)
33 movq %rcx, __percpu 8(%rsi)
45 movq __percpu (%rsi), %rax
46 movq __percpu 8(%rsi), %rdx
H A Dcopy_mc_64.S37 movb (%rsi), %al
40 incq %rsi
52 movq (%rsi), %r8
55 addq $8, %rsi
68 movb (%rsi), %al
71 incq %rsi
H A Dcopy_user_64.S43 0: movb (%rsi),%al
46 inc %rsi
57 2: movq (%rsi),%rax
59 addq $8,%rsi
/linux/tools/arch/x86/lib/
H A Dmemcpy_64.S64 movq 0*8(%rsi), %r8
65 movq 1*8(%rsi), %r9
66 movq 2*8(%rsi), %r10
67 movq 3*8(%rsi), %r11
68 leaq 4*8(%rsi), %rsi
83 addq %rdx, %rsi
97 leaq -4*8(%rsi), %rsi
109 subq %rdx, %rsi
118 movq 0*8(%rsi), %r8
119 movq 1*8(%rsi), %r9
[all …]
/linux/arch/x86/kernel/
H A Dsev_verify_cbit.S25 movq sme_me_mask(%rip), %rsi
26 testq %rsi, %rsi
30 movq sev_status(%rip), %rsi
31 testq %rsi, %rsi
35 movq %cr4, %rsi
38 movq %rsi, %rdx
71 movq %rsi, %cr4
H A Drelocate_kernel_64.S67 movq PTR(VA_CONTROL_PAGE)(%rsi), %r11
90 movq PTR(PA_CONTROL_PAGE)(%rsi), %r8
93 movq PTR(PA_TABLE_PAGE)(%rsi), %r9
96 movq PTR(PA_SWAP_PAGE)(%rsi), %r10
289 movq %rcx, %rsi /* For ever source page do a copy */
290 andq $0xfffffffffffff000, %rsi
293 movq %rsi, %rax
300 movq %rdx, %rsi
305 movq %r10, %rsi
309 lea PAGE_SIZE(%rax), %rsi
/linux/arch/x86/crypto/
H A Dblake2s-core.S71 movd (%rsi,%rax,4),%xmm4
73 movd (%rsi,%rax,4),%xmm5
75 movd (%rsi,%rax,4),%xmm6
77 movd (%rsi,%rax,4),%xmm7
92 movd (%rsi,%rax,4),%xmm5
94 movd (%rsi,%rax,4),%xmm6
96 movd (%rsi,%rax,4),%xmm7
98 movd (%rsi,%rax,4),%xmm4
167 addq $0x40,%rsi
193 vmovdqu (%rsi),%ymm6
[all …]
H A Dsm4-aesni-avx-asm_64.S235 vmovdqu RA0, 0*16(%rsi);
238 vmovdqu RA1, 1*16(%rsi);
240 vmovdqu RA2, 2*16(%rsi);
243 vmovdqu RA3, 3*16(%rsi);
396 vmovdqu RA0, (0 * 16)(%rsi);
397 vmovdqu RA1, (1 * 16)(%rsi);
398 vmovdqu RA2, (2 * 16)(%rsi);
399 vmovdqu RA3, (3 * 16)(%rsi);
400 vmovdqu RB0, (4 * 16)(%rsi);
402 vmovdqu RB1, (5 * 16)(%rsi);
[all …]
H A Daes-gcm-aesni-x86_64.S748 lea 16(KEY), %rsi
751 movdqa (%rsi), TMP0
753 add $16, %rsi
756 movdqa (%rsi), TMP0
770 lea 16(KEY), %rsi
774 movdqa (%rsi), TMP0
775 add $16, %rsi
784 add $16, %rsi
794 add $16, %rsi
930 lea 16(KEY), %rsi
[all …]
H A Dchacha-avx512vl-x86_64.S118 vmovdqu %xmm6,0x00(%rsi)
125 vmovdqu %xmm6,0x10(%rsi)
132 vmovdqu %xmm6,0x20(%rsi)
139 vmovdqu %xmm6,0x30(%rsi)
147 vmovdqu %xmm6,0x40(%rsi)
153 vmovdqu %xmm6,0x50(%rsi)
159 vmovdqu %xmm6,0x60(%rsi)
165 vmovdqu %xmm6,0x70(%rsi)
330 vmovdqu %xmm9,0x00(%rsi)
337 vmovdqu %xmm9,0x10(%rsi)
[all …]
H A Dchacha-avx2-x86_64.S145 vmovdqu %xmm6,0x00(%rsi)
152 vmovdqu %xmm6,0x10(%rsi)
205 mov %rsi,%r11
211 lea (%rdx,%rax),%rsi
219 mov %rsp,%rsi
510 mov %rsi,%r11
516 lea (%rdx,%rax),%rsi
524 mov %rsp,%rsi
1004 mov %rsi,%r11
1006 lea (%rdx,%rax),%rsi
[all …]
H A Dsm4-aesni-avx2-asm_64.S376 vmovdqu RA0, (0 * 32)(%rsi);
377 vmovdqu RA1, (1 * 32)(%rsi);
378 vmovdqu RA2, (2 * 32)(%rsi);
379 vmovdqu RA3, (3 * 32)(%rsi);
380 vmovdqu RB0, (4 * 32)(%rsi);
381 vmovdqu RB1, (5 * 32)(%rsi);
382 vmovdqu RB2, (6 * 32)(%rsi);
383 vmovdqu RB3, (7 * 32)(%rsi);
429 vmovdqu RA0, (0 * 32)(%rsi);
430 vmovdqu RA1, (1 * 32)(%rsi);
[all …]
H A Dchacha-ssse3-x86_64.S141 movdqu %xmm0,0x00(%rsi)
149 movdqu %xmm0,0x10(%rsi)
157 movdqu %xmm0,0x20(%rsi)
165 movdqu %xmm0,0x30(%rsi)
178 mov %rsi,%r11
184 lea (%rdx,%rax),%rsi
192 mov %rsp,%rsi
216 movdqu %xmm0,0x00(%rsi)
774 mov %rsi,%r11
776 lea (%rdx,%rax),%rsi
[all …]
/linux/arch/x86/mm/
H A Dmem_encrypt_boot.S43 movq %rsi, %r11 /* Decrypted area */
48 leaq __enc_copy(%rip), %rsi /* Encryption routine */
54 movq %r11, %rsi /* Decrypted area */
113 movq %rsi, %r11 /* Save decrypted area address */
133 movq %r11, %rsi /* Source - decrypted area */
138 movq %r8, %rsi /* Source - intermediate copy buffer */
/linux/net/sunrpc/auth_gss/
H A Dsvcauth_gss.c98 struct rsi { struct
106 static struct rsi *rsi_update(struct cache_detail *cd, struct rsi *new, struct rsi *old); argument
127 struct rsi *rsii = container_of(ref, struct rsi, h.ref); in rsi_put()
140 struct rsi *item = container_of(a, struct rsi, h); in rsi_match()
141 struct rsi *tmp = container_of(b, struct rsi, h); in rsi_match()
162 struct rsi *new = container_of(cnew, struct rsi, h); in rsi_init()
163 struct rsi *item = container_of(citem, struct rsi, h); in rsi_init()
181 struct rsi *new = container_of(cnew, struct rsi, h); in update_rsi()
182 struct rsi *item = container_of(citem, struct rsi, h); in update_rsi()
216 struct rsi *rsii = container_of(h, struct rsi, h); in rsi_request()
[all …]
/linux/arch/x86/boot/compressed/
H A Didt_handlers_64.S24 pushq %rsi
42 movq ORIG_RAX(%rsp), %rsi
59 popq %rsi
/linux/arch/x86/um/
H A Dsetjmp_64.S25 pop %rsi # Return address, and adjust the stack
29 push %rsi # Make the call/return stack happy
35 movq %rsi,56(%rdi) # Return address
/linux/arch/x86/entry/
H A Dcalling.h70 pushq %rsi /* pt_regs->si */
71 movq 8(%rsp), %rsi /* temporarily store the return address in %rsi */
75 pushq %rsi /* pt_regs->si */
96 pushq %rsi /* return address on top of stack */
144 popq %rsi
438 pushq %rsi
456 popq %rsi
/linux/arch/x86/kernel/cpu/
H A Dvmware.c73 unsigned long out0, rbx, rcx, rdx, rsi, rdi; in vmware_hypercall_slow() local
79 "=d" (rdx), "=S" (rsi), "=D" (rdi) in vmware_hypercall_slow()
91 "=d" (rdx), "=S" (rsi), "=D" (rdi) in vmware_hypercall_slow()
103 "=d" (rdx), "=S" (rsi), "=D" (rdi) in vmware_hypercall_slow()
122 *out4 = rsi; in vmware_hypercall_slow()
517 args.rsi = in4; in vmware_tdx_hypercall()
535 *out4 = args.rsi; in vmware_tdx_hypercall()

123456