10 #ifndef EIGEN_PACKET_MATH_SSE_H 11 #define EIGEN_PACKET_MATH_SSE_H 17 #ifndef EIGEN_CACHEFRIENDLY_PRODUCT_THRESHOLD 18 #define EIGEN_CACHEFRIENDLY_PRODUCT_THRESHOLD 8 21 #ifndef EIGEN_ARCH_DEFAULT_NUMBER_OF_REGISTERS 22 #define EIGEN_ARCH_DEFAULT_NUMBER_OF_REGISTERS (2*sizeof(void*)) 26 #ifndef EIGEN_HAS_SINGLE_INSTRUCTION_MADD 27 #define EIGEN_HAS_SINGLE_INSTRUCTION_MADD 1 31 #if (defined EIGEN_VECTORIZE_AVX) && (EIGEN_COMP_GNUC_STRICT || EIGEN_COMP_MINGW) && (__GXX_ABI_VERSION < 1004) 38 struct eigen_packet_wrapper
51 typedef eigen_packet_wrapper<__m128>
Packet4f;
52 typedef eigen_packet_wrapper<__m128i>
Packet4i;
53 typedef eigen_packet_wrapper<__m128d>
Packet2d;
64 #define vec4f_swizzle1(v,p,q,r,s) \ 65 (_mm_castsi128_ps(_mm_shuffle_epi32( _mm_castps_si128(v), ((s)<<6|(r)<<4|(q)<<2|(p))))) 67 #define vec4i_swizzle1(v,p,q,r,s) \ 68 (_mm_shuffle_epi32( v, ((s)<<6|(r)<<4|(q)<<2|(p)))) 70 #define vec2d_swizzle1(v,p,q) \ 71 (_mm_castsi128_pd(_mm_shuffle_epi32( _mm_castpd_si128(v), ((q*2+1)<<6|(q*2)<<4|(p*2+1)<<2|(p*2))))) 73 #define vec4f_swizzle2(a,b,p,q,r,s) \ 74 (_mm_shuffle_ps( (a), (b), ((s)<<6|(r)<<4|(q)<<2|(p)))) 76 #define vec4i_swizzle2(a,b,p,q,r,s) \ 77 (_mm_castps_si128( (_mm_shuffle_ps( _mm_castsi128_ps(a), _mm_castsi128_ps(b), ((s)<<6|(r)<<4|(q)<<2|(p)))))) 79 #define _EIGEN_DECLARE_CONST_Packet4f(NAME,X) \ 80 const Packet4f p4f_##NAME = pset1<Packet4f>(X) 82 #define _EIGEN_DECLARE_CONST_Packet2d(NAME,X) \ 83 const Packet2d p2d_##NAME = pset1<Packet2d>(X) 85 #define _EIGEN_DECLARE_CONST_Packet4f_FROM_INT(NAME,X) \ 86 const Packet4f p4f_##NAME = _mm_castsi128_ps(pset1<Packet4i>(X)) 88 #define _EIGEN_DECLARE_CONST_Packet4i(NAME,X) \ 89 const Packet4i p4i_##NAME = pset1<Packet4i>(X) 94 #ifndef EIGEN_VECTORIZE_AVX 95 template<>
struct packet_traits<float> : default_packet_traits
115 #ifdef EIGEN_VECTORIZE_SSE4_1 139 #ifdef EIGEN_VECTORIZE_SSE4_1 165 #ifndef EIGEN_VECTORIZE_AVX 167 template<>
struct scalar_div_cost<double,true> {
enum { value = 8 }; };
170 #if EIGEN_COMP_MSVC==1500 188 #if EIGEN_COMP_GNUC_STRICT && (!defined __AVX__) 208 const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x80000000,0x80000000,0x80000000,0x80000000));
209 return _mm_xor_ps(a,mask);
213 const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0x0,0x80000000,0x0,0x80000000));
214 return _mm_xor_pd(a,mask);
229 #ifdef EIGEN_VECTORIZE_SSE4_1 230 return _mm_mullo_epi32(a,
b);
249 template<>
EIGEN_STRONG_INLINE Packet4f
pmadd(
const Packet4f& a,
const Packet4f& b,
const Packet4f& c) {
return _mm_fmadd_ps(a,b,c); }
250 template<>
EIGEN_STRONG_INLINE Packet2d
pmadd(
const Packet2d& a,
const Packet2d& b,
const Packet2d& c) {
return _mm_fmadd_pd(a,b,c); }
257 #ifdef EIGEN_VECTORIZE_SSE4_1 258 return _mm_min_epi32(a,b);
261 Packet4i mask = _mm_cmplt_epi32(a,b);
262 return _mm_or_si128(_mm_and_si128(mask,a),_mm_andnot_si128(mask,b));
270 #ifdef EIGEN_VECTORIZE_SSE4_1 271 return _mm_max_epi32(a,b);
274 Packet4i mask = _mm_cmpgt_epi32(a,b);
275 return _mm_or_si128(_mm_and_si128(mask,a),_mm_andnot_si128(mask,b));
279 #ifdef EIGEN_VECTORIZE_SSE4_1 313 #if (EIGEN_COMP_MSVC==1600) 316 __m128 res = _mm_loadl_pi(_mm_set1_ps(0.0
f), (
const __m64*)(from));
317 res = _mm_loadh_pi(res, (
const __m64*)(from+2));
320 return _mm_loadu_ps(from);
329 return _mm_loadu_ps(from);
336 return _mm_loadu_pd(from);
341 return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
347 return vec4f_swizzle1(_mm_castpd_ps(_mm_load_sd(reinterpret_cast<const double*>(from))), 0, 0, 1, 1);
354 tmp = _mm_loadl_epi64(reinterpret_cast<const __m128i*>(from));
368 return _mm_set_ps(from[3*stride], from[2*stride], from[1*stride], from[0*stride]);
372 return _mm_set_pd(from[1*stride], from[0*stride]);
376 return _mm_set_epi32(from[3*stride], from[2*stride], from[1*stride], from[0*stride]);
381 to[stride*0] = _mm_cvtss_f32(from);
382 to[stride*1] = _mm_cvtss_f32(_mm_shuffle_ps(from, from, 1));
383 to[stride*2] = _mm_cvtss_f32(_mm_shuffle_ps(from, from, 2));
384 to[stride*3] = _mm_cvtss_f32(_mm_shuffle_ps(from, from, 3));
388 to[stride*0] = _mm_cvtsd_f64(from);
389 to[stride*1] = _mm_cvtsd_f64(_mm_shuffle_pd(from, from, 1));
393 to[stride*0] = _mm_cvtsi128_si32(from);
394 to[stride*1] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 1));
395 to[stride*2] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 2));
396 to[stride*3] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 3));
402 Packet4f pa = _mm_set_ss(a);
408 Packet2d pa = _mm_set_sd(a);
412 #ifndef EIGEN_VECTORIZE_AVX 418 #if EIGEN_COMP_MSVC_STRICT && EIGEN_OS_WIN64 424 #elif EIGEN_COMP_MSVC_STRICT 436 {
return _mm_shuffle_ps(a,a,0x1B); }
438 {
return _mm_shuffle_pd(a,a,0x1); }
440 {
return _mm_shuffle_epi32(a,0x1B); }
444 const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x7FFFFFFF,0x7FFFFFFF,0x7FFFFFFF,0x7FFFFFFF));
445 return _mm_and_ps(a,mask);
449 const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0xFFFFFFFF,0x7FFFFFFF,0xFFFFFFFF,0x7FFFFFFF));
450 return _mm_and_pd(a,mask);
454 #ifdef EIGEN_VECTORIZE_SSSE3 455 return _mm_abs_epi32(a);
457 Packet4i aux = _mm_srai_epi32(a,31);
458 return _mm_sub_epi32(_mm_xor_si128(a,aux),aux);
466 Packet4f& a0, Packet4f& a1, Packet4f& a2, Packet4f& a3)
476 Packet2d& a0, Packet2d& a1, Packet2d& a2, Packet2d& a3)
478 #ifdef EIGEN_VECTORIZE_SSE3 479 a0 = _mm_loaddup_pd(a+0);
480 a1 = _mm_loaddup_pd(a+1);
481 a2 = _mm_loaddup_pd(a+2);
482 a3 = _mm_loaddup_pd(a+3);
496 vecs[1] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0x55));
497 vecs[2] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0xAA));
498 vecs[3] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0xFF));
499 vecs[0] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0x00));
502 #ifdef EIGEN_VECTORIZE_SSE3 505 return _mm_hadd_ps(_mm_hadd_ps(vecs[0], vecs[1]),_mm_hadd_ps(vecs[2], vecs[3]));
510 return _mm_hadd_pd(vecs[0], vecs[1]);
516 Packet4f tmp0, tmp1, tmp2;
517 tmp0 = _mm_unpacklo_ps(vecs[0], vecs[1]);
518 tmp1 = _mm_unpackhi_ps(vecs[0], vecs[1]);
519 tmp2 = _mm_unpackhi_ps(vecs[2], vecs[3]);
520 tmp0 = _mm_add_ps(tmp0, tmp1);
521 tmp1 = _mm_unpacklo_ps(vecs[2], vecs[3]);
522 tmp1 = _mm_add_ps(tmp1, tmp2);
523 tmp2 = _mm_movehl_ps(tmp1, tmp0);
524 tmp0 = _mm_movelh_ps(tmp0, tmp1);
525 return _mm_add_ps(tmp0, tmp2);
530 return _mm_add_pd(_mm_unpacklo_pd(vecs[0], vecs[1]), _mm_unpackhi_pd(vecs[0], vecs[1]));
542 Packet4f tmp = _mm_add_ps(a, _mm_movehl_ps(a,a));
558 #ifdef EIGEN_VECTORIZE_SSSE3 561 return _mm_hadd_epi32(_mm_hadd_epi32(vecs[0], vecs[1]),_mm_hadd_epi32(vecs[2], vecs[3]));
565 Packet4i tmp0 = _mm_hadd_epi32(a,a);
571 Packet4i tmp = _mm_add_epi32(a, _mm_unpackhi_epi64(a,a));
577 Packet4i tmp0, tmp1, tmp2;
578 tmp0 = _mm_unpacklo_epi32(vecs[0], vecs[1]);
579 tmp1 = _mm_unpackhi_epi32(vecs[0], vecs[1]);
580 tmp2 = _mm_unpackhi_epi32(vecs[2], vecs[3]);
581 tmp0 = _mm_add_epi32(tmp0, tmp1);
582 tmp1 = _mm_unpacklo_epi32(vecs[2], vecs[3]);
583 tmp1 = _mm_add_epi32(tmp1, tmp2);
584 tmp2 = _mm_unpacklo_epi64(tmp0, tmp1);
585 tmp0 = _mm_unpackhi_epi64(tmp0, tmp1);
586 return _mm_add_epi32(tmp0, tmp2);
594 Packet4f tmp = _mm_mul_ps(a, _mm_movehl_ps(a,a));
608 return (aux[0] * aux[1]) * (aux[2] * aux[3]);;
614 Packet4f tmp = _mm_min_ps(a, _mm_movehl_ps(a,a));
623 #ifdef EIGEN_VECTORIZE_SSE4_1 624 Packet4i tmp = _mm_min_epi32(a, _mm_shuffle_epi32(a, _MM_SHUFFLE(0,0,3,2)));
631 int aux0 = aux[0]<aux[1] ? aux[0] : aux[1];
632 int aux2 = aux[2]<aux[3] ? aux[2] : aux[3];
633 return aux0<aux2 ? aux0 : aux2;
634 #endif // EIGEN_VECTORIZE_SSE4_1 640 Packet4f tmp = _mm_max_ps(a, _mm_movehl_ps(a,a));
649 #ifdef EIGEN_VECTORIZE_SSE4_1 650 Packet4i tmp = _mm_max_epi32(a, _mm_shuffle_epi32(a, _MM_SHUFFLE(0,0,3,2)));
657 int aux0 = aux[0]>aux[1] ? aux[0] : aux[1];
658 int aux2 = aux[2]>aux[3] ? aux[2] : aux[3];
659 return aux0>aux2 ? aux0 : aux2;
660 #endif // EIGEN_VECTORIZE_SSE4_1 678 #ifdef EIGEN_VECTORIZE_SSSE3 686 first = _mm_castsi128_ps(_mm_alignr_epi8(_mm_castps_si128(second), _mm_castps_si128(first), Offset*4));
696 first = _mm_alignr_epi8(second,first, Offset*4);
706 first = _mm_castsi128_pd(_mm_alignr_epi8(_mm_castpd_si128(second), _mm_castpd_si128(first), 8));
718 first = _mm_move_ss(first,second);
719 first = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(first),0x39));
723 first = _mm_movehl_ps(first,first);
724 first = _mm_movelh_ps(first,second);
728 first = _mm_move_ss(first,second);
729 first = _mm_shuffle_ps(first,second,0x93);
741 first = _mm_castps_si128(_mm_move_ss(_mm_castsi128_ps(first),_mm_castsi128_ps(second)));
742 first = _mm_shuffle_epi32(first,0x39);
746 first = _mm_castps_si128(_mm_movehl_ps(_mm_castsi128_ps(first),_mm_castsi128_ps(first)));
747 first = _mm_castps_si128(_mm_movelh_ps(_mm_castsi128_ps(first),_mm_castsi128_ps(second)));
751 first = _mm_castps_si128(_mm_move_ss(_mm_castsi128_ps(first),_mm_castsi128_ps(second)));
752 first = _mm_castps_si128(_mm_shuffle_ps(_mm_castsi128_ps(first),_mm_castsi128_ps(second),0x93));
764 first = _mm_castps_pd(_mm_movehl_ps(_mm_castpd_ps(first),_mm_castpd_ps(first)));
765 first = _mm_castps_pd(_mm_movelh_ps(_mm_castpd_ps(first),_mm_castpd_ps(second)));
771 EIGEN_DEVICE_FUNC
inline void 776 EIGEN_DEVICE_FUNC
inline void 778 __m128d tmp = _mm_unpackhi_pd(kernel.
packet[0], kernel.
packet[1]);
783 EIGEN_DEVICE_FUNC
inline void 785 __m128i T0 = _mm_unpacklo_epi32(kernel.
packet[0], kernel.
packet[1]);
786 __m128i T1 = _mm_unpacklo_epi32(kernel.
packet[2], kernel.
packet[3]);
787 __m128i T2 = _mm_unpackhi_epi32(kernel.
packet[0], kernel.
packet[1]);
788 __m128i T3 = _mm_unpackhi_epi32(kernel.
packet[2], kernel.
packet[3]);
790 kernel.
packet[0] = _mm_unpacklo_epi64(T0, T1);
791 kernel.
packet[1] = _mm_unpackhi_epi64(T0, T1);
792 kernel.
packet[2] = _mm_unpacklo_epi64(T2, T3);
793 kernel.
packet[3] = _mm_unpackhi_epi64(T2, T3);
797 const __m128i zero = _mm_setzero_si128();
798 const __m128i select = _mm_set_epi32(ifPacket.
select[3], ifPacket.
select[2], ifPacket.
select[1], ifPacket.
select[0]);
799 __m128i false_mask = _mm_cmpeq_epi32(select, zero);
800 #ifdef EIGEN_VECTORIZE_SSE4_1 801 return _mm_blendv_epi8(thenPacket, elsePacket, false_mask);
803 return _mm_or_si128(_mm_andnot_si128(false_mask, thenPacket), _mm_and_si128(false_mask, elsePacket));
807 const __m128 zero = _mm_setzero_ps();
809 __m128 false_mask = _mm_cmpeq_ps(select, zero);
810 #ifdef EIGEN_VECTORIZE_SSE4_1 811 return _mm_blendv_ps(thenPacket, elsePacket, false_mask);
813 return _mm_or_ps(_mm_andnot_ps(false_mask, thenPacket), _mm_and_ps(false_mask, elsePacket));
817 const __m128d zero = _mm_setzero_pd();
818 const __m128d select = _mm_set_pd(ifPacket.
select[1], ifPacket.
select[0]);
819 __m128d false_mask = _mm_cmpeq_pd(select, zero);
820 #ifdef EIGEN_VECTORIZE_SSE4_1 821 return _mm_blendv_pd(thenPacket, elsePacket, false_mask);
823 return _mm_or_pd(_mm_andnot_pd(false_mask, thenPacket), _mm_and_pd(false_mask, elsePacket));
829 #ifdef EIGEN_VECTORIZE_SSE4_1 832 return _mm_move_ss(a, _mm_load_ss(&b));
838 #ifdef EIGEN_VECTORIZE_SSE4_1 841 return _mm_move_sd(a, _mm_load_sd(&b));
847 #ifdef EIGEN_VECTORIZE_SSE4_1 850 const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x0,0x0,0x0,0xFFFFFFFF));
851 return _mm_or_ps(_mm_andnot_ps(mask, a), _mm_and_ps(mask,
pset1<Packet4f>(b)));
857 #ifdef EIGEN_VECTORIZE_SSE4_1 860 const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0x0,0x0,0xFFFFFFFF,0xFFFFFFFF));
861 return _mm_or_pd(_mm_andnot_pd(mask, a), _mm_and_pd(mask,
pset1<Packet2d>(b)));
868 return ::fmaf(a,b,c);
879 #endif // EIGEN_PACKET_MATH_SSE_H EIGEN_STRONG_INLINE Packet4i ploaddup< Packet4i >(const int *from)
EIGEN_DEVICE_FUNC void pscatter< int, Packet4i >(int *to, const Packet4i &from, Index stride)
#define EIGEN_ALWAYS_INLINE
EIGEN_STRONG_INLINE Packet4f pxor< Packet4f >(const Packet4f &a, const Packet4f &b)
#define EIGEN_STRONG_INLINE
EIGEN_STRONG_INLINE Packet4f pround< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE Packet2d plset< Packet2d >(const double &a)
EIGEN_STRONG_INLINE int pfirst< Packet4i >(const Packet4i &a)
EIGEN_DEVICE_FUNC void pscatter< float, Packet4f >(float *to, const Packet4f &from, Index stride)
EIGEN_STRONG_INLINE double predux_min< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE Packet4i pload< Packet4i >(const int *from)
EIGEN_STRONG_INLINE double predux_max< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE Packet4cf pinsertfirst(const Packet4cf &a, std::complex< float > b)
EIGEN_STRONG_INLINE Packet4i plset< Packet4i >(const int &a)
#define EIGEN_DEBUG_UNALIGNED_LOAD
EIGEN_STRONG_INLINE Packet2d pceil< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE float pfirst< Packet4f >(const Packet4f &a)
static int f(const TensorMap< Tensor< int, 3 > > &tensor)
EIGEN_STRONG_INLINE Packet4f ploaddup< Packet4f >(const float *from)
EIGEN_STRONG_INLINE void pbroadcast4< Packet4f >(const float *a, Packet4f &a0, Packet4f &a1, Packet4f &a2, Packet4f &a3)
EIGEN_STRONG_INLINE int predux< Packet4i >(const Packet4i &a)
#define EIGEN_DEBUG_ALIGNED_STORE
static EIGEN_STRONG_INLINE void run(Packet4i &first, const Packet4i &second)
EIGEN_STRONG_INLINE Packet4i ploadu< Packet4i >(const int *from)
static constexpr size_t size(Tuple< Args... > &)
Provides access to the number of elements in a tuple as a compile-time constant expression.
EIGEN_STRONG_INLINE float predux_max< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE void pstore1< Packet2d >(double *to, const double &a)
static EIGEN_STRONG_INLINE void run(Packet4f &first, const Packet4f &second)
#define EIGEN_DEBUG_UNALIGNED_STORE
EIGEN_STRONG_INLINE void prefetch< float >(const float *addr)
EIGEN_STRONG_INLINE int predux_min< Packet4i >(const Packet4i &a)
EIGEN_STRONG_INLINE void pstoreu< double >(double *to, const Packet4d &from)
#define EIGEN_DEBUG_ALIGNED_LOAD
#define vec4i_swizzle1(v, p, q, r, s)
EIGEN_STRONG_INLINE Packet2d padd< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE Packet2d ploaddup< Packet2d >(const double *from)
EIGEN_STRONG_INLINE double predux_mul< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE Packet2d preduxp< Packet2d >(const Packet2d *vecs)
EIGEN_DEVICE_FUNC Packet4i pgather< int, Packet4i >(const int *from, Index stride)
EIGEN_DEVICE_FUNC Packet padd(const Packet &a, const Packet &b)
EIGEN_STRONG_INLINE Packet2d pdiv< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE Packet4i pmul< Packet4i >(const Packet4i &a, const Packet4i &b)
#define vec2d_swizzle1(v, p, q)
EIGEN_STRONG_INLINE Packet2d pandnot< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE void pstore< double >(double *to, const Packet4d &from)
EIGEN_STRONG_INLINE Packet4f pceil< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE Packet4f pdiv< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE Packet4f pmin< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE Packet2d pmin< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE void prefetch< int >(const int *addr)
EIGEN_STRONG_INLINE Packet4i pand< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE int predux_max< Packet4i >(const Packet4i &a)
EIGEN_STRONG_INLINE Packet4i pmin< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE Packet4f ploadu< Packet4f >(const float *from)
EIGEN_STRONG_INLINE void ptranspose(PacketBlock< Packet2cf, 2 > &kernel)
EIGEN_STRONG_INLINE Packet4i pxor< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE void pstoreu< int >(int *to, const Packet4i &from)
#define vec4f_swizzle1(v, p, q, r, s)
EIGEN_STRONG_INLINE void pstore< float >(float *to, const Packet4f &from)
EIGEN_DEFAULT_DENSE_INDEX_TYPE Index
The Index type as used for the API.
EIGEN_DEVICE_FUNC void pscatter< double, Packet2d >(double *to, const Packet2d &from, Index stride)
EIGEN_STRONG_INLINE Packet4f por< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE Packet2d pmax< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE Packet4i preduxp< Packet4i >(const Packet4i *vecs)
EIGEN_DEVICE_FUNC unpacket_traits< Packet >::type pfirst(const Packet &a)
EIGEN_STRONG_INLINE Packet4f pandnot< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE void pbroadcast4< Packet2d >(const double *a, Packet2d &a0, Packet2d &a1, Packet2d &a2, Packet2d &a3)
EIGEN_STRONG_INLINE float predux< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE Packet2d pand< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE Packet2d ploadu< Packet2d >(const double *from)
EIGEN_STRONG_INLINE Packet2cf pconj(const Packet2cf &a)
EIGEN_STRONG_INLINE float predux_min< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE Packet2d pload< Packet2d >(const double *from)
EIGEN_STRONG_INLINE Packet2d psub< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE void pstoreu< float >(float *to, const Packet4f &from)
EIGEN_STRONG_INLINE int predux_mul< Packet4i >(const Packet4i &a)
EIGEN_STRONG_INLINE Packet4f pmul< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_DEVICE_FUNC void pstore(Scalar *to, const Packet &from)
EIGEN_STRONG_INLINE double predux< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE Packet4f pload< Packet4f >(const float *from)
EIGEN_STRONG_INLINE Packet4i pmax< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE Packet2d pxor< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE Packet2cf pnegate(const Packet2cf &a)
EIGEN_STRONG_INLINE Packet4i por< Packet4i >(const Packet4i &a, const Packet4i &b)
const mpreal fma(const mpreal &v1, const mpreal &v2, const mpreal &v3, mp_rnd_t rnd_mode=mpreal::get_default_rnd())
EIGEN_STRONG_INLINE Packet4i padd< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE void pstore1< Packet4f >(float *to, const float &a)
EIGEN_STRONG_INLINE Packet4f padd< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE void pstore< int >(int *to, const Packet4i &from)
EIGEN_STRONG_INLINE Packet4i pandnot< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE Packet4f pfloor< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE Packet4f pmadd(const Packet4f &a, const Packet4f &b, const Packet4f &c)
EIGEN_DEVICE_FUNC Packet psub(const Packet &a, const Packet &b)
EIGEN_STRONG_INLINE Packet4f preduxp< Packet4f >(const Packet4f *vecs)
EIGEN_STRONG_INLINE Packet2d pround< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE Packet2d pfloor< Packet2d >(const Packet2d &a)
EIGEN_STRONG_INLINE Packet4cf pinsertlast(const Packet4cf &a, std::complex< float > b)
EIGEN_STRONG_INLINE Packet2d pset1< Packet2d >(const double &from)
EIGEN_STRONG_INLINE Packet4f plset< Packet4f >(const float &a)
EIGEN_DEVICE_FUNC Packet2d pgather< double, Packet2d >(const double *from, Index stride)
EIGEN_STRONG_INLINE Packet2d pmul< Packet2d >(const Packet2d &a, const Packet2d &b)
static EIGEN_STRONG_INLINE void run(Packet2d &first, const Packet2d &second)
void run(Expr &expr, Dev &dev)
EIGEN_STRONG_INLINE Packet4f pmax< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE Packet4f pset1< Packet4f >(const float &from)
EIGEN_DEVICE_FUNC const Scalar & b
EIGEN_STRONG_INLINE void punpackp(Packet4f *vecs)
EIGEN_STRONG_INLINE double pfirst< Packet2d >(const Packet2d &a)
EIGEN_DEVICE_FUNC Packet pmul(const Packet &a, const Packet &b)
EIGEN_DEVICE_FUNC Packet4f pgather< float, Packet4f >(const float *from, Index stride)
EIGEN_STRONG_INLINE void prefetch< double >(const double *addr)
EIGEN_STRONG_INLINE Packet4i psub< Packet4i >(const Packet4i &a, const Packet4i &b)
EIGEN_STRONG_INLINE Packet4f psub< Packet4f >(const Packet4f &a, const Packet4f &b)
EIGEN_STRONG_INLINE Packet4i pblend(const Selector< 4 > &ifPacket, const Packet4i &thenPacket, const Packet4i &elsePacket)
#define vec4i_swizzle2(a, b, p, q, r, s)
EIGEN_STRONG_INLINE Packet4i pset1< Packet4i >(const int &from)
EIGEN_STRONG_INLINE Packet2cf preverse(const Packet2cf &a)
EIGEN_STRONG_INLINE Packet2d por< Packet2d >(const Packet2d &a, const Packet2d &b)
EIGEN_STRONG_INLINE float predux_mul< Packet4f >(const Packet4f &a)
EIGEN_STRONG_INLINE Packet4f pabs(const Packet4f &a)
EIGEN_STRONG_INLINE Packet4f pand< Packet4f >(const Packet4f &a, const Packet4f &b)