< prev index next >
src/cpu/x86/vm/stubGenerator_x86_64.cpp
Print this page
@@ -3234,15 +3234,10 @@
}
#ifdef _WIN64
// on win64, fill len_reg from stack position
__ movl(len_reg, len_mem);
- // save the xmm registers which must be preserved 6-15
- __ subptr(rsp, -rsp_after_call_off * wordSize);
- for (int i = 6; i <= XMM_REG_NUM_KEY_LAST; i++) {
- __ movdqu(xmm_save(i), as_XMMRegister(i));
- }
#else
__ push(len_reg); // Save
#endif
const XMMRegister xmm_key_shuf_mask = xmm_temp; // used temporarily to swap key bytes up front
@@ -3279,14 +3274,10 @@
__ BIND(L_exit);
__ movdqu(Address(rvec, 0), xmm_result); // final value of r stored in rvec of CipherBlockChaining object
#ifdef _WIN64
- // restore xmm regs belonging to calling function
- for (int i = 6; i <= XMM_REG_NUM_KEY_LAST; i++) {
- __ movdqu(as_XMMRegister(i), xmm_save(i));
- }
__ movl(rax, len_mem);
#else
__ pop(rax); // return length
#endif
__ leave(); // required for proper stackwalking of RuntimeStub frame
@@ -3444,15 +3435,10 @@
}
#ifdef _WIN64
// on win64, fill len_reg from stack position
__ movl(len_reg, len_mem);
- // save the xmm registers which must be preserved 6-15
- __ subptr(rsp, -rsp_after_call_off * wordSize);
- for (int i = 6; i <= XMM_REG_NUM_KEY_LAST; i++) {
- __ movdqu(xmm_save(i), as_XMMRegister(i));
- }
#else
__ push(len_reg); // Save
#endif
__ push(rbx);
// the java expanded key ordering is rotated one position from what we want
@@ -3642,14 +3628,10 @@
__ BIND(L_exit);
__ movdqu(Address(rvec, 0), xmm_prev_block_cipher); // final value of r stored in rvec of CipherBlockChaining object
__ pop(rbx);
#ifdef _WIN64
- // restore regs belonging to calling function
- for (int i = 6; i <= XMM_REG_NUM_KEY_LAST; i++) {
- __ movdqu(as_XMMRegister(i), xmm_save(i));
- }
__ movl(rax, len_mem);
#else
__ pop(rax); // return length
#endif
__ leave(); // required for proper stackwalking of RuntimeStub frame
@@ -3697,29 +3679,16 @@
const XMMRegister msg3 = xmm6;
const XMMRegister shuf_mask = xmm7;
__ enter();
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
-#endif
-
__ subptr(rsp, 4 * wordSize);
__ fast_sha1(abcd, e0, e1, msg0, msg1, msg2, msg3, shuf_mask,
buf, state, ofs, limit, rsp, multi_block);
__ addptr(rsp, 4 * wordSize);
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
-#endif
__ leave();
__ ret(0);
return start;
}
@@ -3773,26 +3742,10 @@
const XMMRegister msgtmp4 = xmm7;
const XMMRegister shuf_mask = xmm8;
__ enter();
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 6 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
- __ movdqu(Address(rsp, 4 * wordSize), xmm8);
-
- if (!VM_Version::supports_sha() && VM_Version::supports_avx2()) {
- __ subptr(rsp, 10 * wordSize);
- __ movdqu(Address(rsp, 0), xmm9);
- __ movdqu(Address(rsp, 2 * wordSize), xmm10);
- __ movdqu(Address(rsp, 4 * wordSize), xmm11);
- __ movdqu(Address(rsp, 6 * wordSize), xmm12);
- __ movdqu(Address(rsp, 8 * wordSize), xmm13);
- }
-#endif
__ subptr(rsp, 4 * wordSize);
if (VM_Version::supports_sha()) {
__ fast_sha256(msg, state0, state1, msgtmp0, msgtmp1, msgtmp2, msgtmp3, msgtmp4,
@@ -3800,25 +3753,11 @@
} else if (VM_Version::supports_avx2()) {
__ sha256_AVX2(msg, state0, state1, msgtmp0, msgtmp1, msgtmp2, msgtmp3, msgtmp4,
buf, state, ofs, limit, rsp, multi_block, shuf_mask);
}
__ addptr(rsp, 4 * wordSize);
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- if (!VM_Version::supports_sha() && VM_Version::supports_avx2()) {
- __ movdqu(xmm9, Address(rsp, 0));
- __ movdqu(xmm10, Address(rsp, 2 * wordSize));
- __ movdqu(xmm11, Address(rsp, 4 * wordSize));
- __ movdqu(xmm12, Address(rsp, 6 * wordSize));
- __ movdqu(xmm13, Address(rsp, 8 * wordSize));
- __ addptr(rsp, 10 * wordSize);
- }
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ movdqu(xmm8, Address(rsp, 4 * wordSize));
- __ addptr(rsp, 6 * wordSize);
-#endif
+
__ leave();
__ ret(0);
return start;
}
@@ -3915,22 +3854,18 @@
__ movl(rax, 0xffff);
__ kmovql(k1, rax);
}
#ifdef _WIN64
- // save the xmm registers which must be preserved 6-14
- const int XMM_REG_NUM_KEY_LAST = 14;
- __ subptr(rsp, -rsp_after_call_off * wordSize);
- for (int i = 6; i <= XMM_REG_NUM_KEY_LAST; i++) {
- __ movdqu(xmm_save(i), as_XMMRegister(i));
- }
-
- const Address r13_save(rbp, rdi_off * wordSize);
- const Address r14_save(rbp, rsi_off * wordSize);
-
- __ movptr(r13_save, r13);
- __ movptr(r14_save, r14);
+ // allocate spill slots for r13, r14
+ enum {
+ saved_r13_offset,
+ saved_r14_offset
+ };
+ __ subptr(rsp, 2 * wordSize);
+ __ movptr(Address(rsp, saved_r13_offset * wordSize), r13);
+ __ movptr(Address(rsp, saved_r14_offset * wordSize), r14);
// on win64, fill len_reg from stack position
__ movl(len_reg, len_mem);
__ movptr(saved_encCounter_start, saved_encCounter_mem);
__ movptr(used_addr, used_mem);
@@ -4128,17 +4063,14 @@
__ BIND(L_exit);
__ pshufb(xmm_curr_counter, xmm_counter_shuf_mask); //counter is shuffled back.
__ movdqu(Address(counter, 0), xmm_curr_counter); //save counter back
__ pop(rbx); // pop the saved RBX.
#ifdef _WIN64
- // restore regs belonging to calling function
- for (int i = 6; i <= XMM_REG_NUM_KEY_LAST; i++) {
- __ movdqu(as_XMMRegister(i), xmm_save(i));
- }
__ movl(rax, len_mem);
- __ movptr(r13, r13_save);
- __ movptr(r14, r14_save);
+ __ movptr(r13, Address(rsp, saved_r13_offset * wordSize));
+ __ movptr(r14, Address(rsp, saved_r14_offset * wordSize));
+ __ addptr(rsp, 2 * wordSize);
#else
__ pop(rax); // return 'len'
#endif
__ leave(); // required for proper stackwalking of RuntimeStub frame
__ ret(0);
@@ -4175,14 +4107,10 @@
const Register state = c_rarg0;
const Register subkeyH = c_rarg1;
const Register data = c_rarg2;
const Register blocks = c_rarg3;
-#ifdef _WIN64
- const int XMM_REG_LAST = 10;
-#endif
-
const XMMRegister xmm_temp0 = xmm0;
const XMMRegister xmm_temp1 = xmm1;
const XMMRegister xmm_temp2 = xmm2;
const XMMRegister xmm_temp3 = xmm3;
const XMMRegister xmm_temp4 = xmm4;
@@ -4201,18 +4129,10 @@
if (VM_Version::supports_avx512vlbw()) {
__ movl(rax, 0xffff);
__ kmovql(k1, rax);
}
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-10
- __ subptr(rsp, -rsp_after_call_off * wordSize);
- for (int i = 6; i <= XMM_REG_LAST; i++) {
- __ movdqu(xmm_save(i), as_XMMRegister(i));
- }
-#endif
-
__ movdqu(xmm_temp10, ExternalAddress(StubRoutines::x86::ghash_long_swap_mask_addr()));
__ movdqu(xmm_temp0, Address(state, 0));
__ pshufb(xmm_temp0, xmm_temp10);
@@ -4308,16 +4228,10 @@
__ BIND(L_exit);
__ pshufb(xmm_temp6, xmm_temp10); // Byte swap 16-byte result
__ movdqu(Address(state, 0), xmm_temp6); // store the result
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- for (int i = 6; i <= XMM_REG_LAST; i++) {
- __ movdqu(as_XMMRegister(i), xmm_save(i));
- }
-#endif
__ leave();
__ ret(0);
return start;
}
@@ -4650,25 +4564,12 @@
const Register tmp = r11;
BLOCK_COMMENT("Entry:");
__ enter(); // required for proper stackwalking of RuntimeStub frame
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
-#endif
__ fast_exp(x0, x1, x2, x3, x4, x5, x6, x7, rax, rcx, rdx, tmp);
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
-#endif
-
__ leave(); // required for proper stackwalking of RuntimeStub frame
__ ret(0);
return start;
@@ -4691,25 +4592,12 @@
const Register tmp2 = r8;
BLOCK_COMMENT("Entry:");
__ enter(); // required for proper stackwalking of RuntimeStub frame
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
-#endif
__ fast_log(x0, x1, x2, x3, x4, x5, x6, x7, rax, rcx, rdx, tmp1, tmp2);
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
-#endif
-
__ leave(); // required for proper stackwalking of RuntimeStub frame
__ ret(0);
return start;
@@ -4731,25 +4619,12 @@
const Register tmp = r11;
BLOCK_COMMENT("Entry:");
__ enter(); // required for proper stackwalking of RuntimeStub frame
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
-#endif
__ fast_log10(x0, x1, x2, x3, x4, x5, x6, x7, rax, rcx, rdx, tmp);
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
-#endif
-
__ leave(); // required for proper stackwalking of RuntimeStub frame
__ ret(0);
return start;
@@ -4774,25 +4649,12 @@
const Register tmp4 = r11;
BLOCK_COMMENT("Entry:");
__ enter(); // required for proper stackwalking of RuntimeStub frame
-#ifdef _WIN64
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
-#endif
__ fast_pow(x0, x1, x2, x3, x4, x5, x6, x7, rax, rcx, rdx, tmp1, tmp2, tmp3, tmp4);
-#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
-#endif
-
__ leave(); // required for proper stackwalking of RuntimeStub frame
__ ret(0);
return start;
@@ -4820,22 +4682,14 @@
__ enter(); // required for proper stackwalking of RuntimeStub frame
#ifdef _WIN64
__ push(rsi);
__ push(rdi);
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
#endif
__ fast_sin(x0, x1, x2, x3, x4, x5, x6, x7, rax, rbx, rcx, rdx, tmp1, tmp2, tmp3, tmp4);
#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
__ pop(rdi);
__ pop(rsi);
#endif
__ leave(); // required for proper stackwalking of RuntimeStub frame
@@ -4867,22 +4721,14 @@
__ enter(); // required for proper stackwalking of RuntimeStub frame
#ifdef _WIN64
__ push(rsi);
__ push(rdi);
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
#endif
__ fast_cos(x0, x1, x2, x3, x4, x5, x6, x7, rax, rcx, rdx, tmp1, tmp2, tmp3, tmp4);
#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
__ pop(rdi);
__ pop(rsi);
#endif
__ leave(); // required for proper stackwalking of RuntimeStub frame
@@ -4914,22 +4760,14 @@
__ enter(); // required for proper stackwalking of RuntimeStub frame
#ifdef _WIN64
__ push(rsi);
__ push(rdi);
- // save the xmm registers which must be preserved 6-7
- __ subptr(rsp, 4 * wordSize);
- __ movdqu(Address(rsp, 0), xmm6);
- __ movdqu(Address(rsp, 2 * wordSize), xmm7);
#endif
__ fast_tan(x0, x1, x2, x3, x4, x5, x6, x7, rax, rcx, rdx, tmp1, tmp2, tmp3, tmp4);
#ifdef _WIN64
- // restore xmm regs belonging to calling function
- __ movdqu(xmm6, Address(rsp, 0));
- __ movdqu(xmm7, Address(rsp, 2 * wordSize));
- __ addptr(rsp, 4 * wordSize);
__ pop(rdi);
__ pop(rsi);
#endif
__ leave(); // required for proper stackwalking of RuntimeStub frame
< prev index next >