< prev index next >

src/cpu/x86/vm/assembler_x86.hpp

Print this page

        

*** 2139,2148 **** --- 2139,2153 ---- // penalty if legacy SSE instructions are encoded using VEX prefix because // they always clear upper 128 bits. It should be used before calling // runtime code and native libraries. void vzeroupper(); + // AVX support for vectorized conditional move (double). The following two instructions used only coupled. + void cmppd(XMMRegister dst, XMMRegister nds, XMMRegister src, int cop, int vector_len); + void vpblendd(XMMRegister dst, XMMRegister nds, XMMRegister src1, XMMRegister src2, int vector_len); + + protected: // Next instructions require address alignment 16 bytes SSE mode. // They should be called only from corresponding MacroAssembler instructions. void andpd(XMMRegister dst, Address src); void andps(XMMRegister dst, Address src);
< prev index next >