10 #ifndef EIGEN_GENERAL_MATRIX_MATRIX_H 11 #define EIGEN_GENERAL_MATRIX_MATRIX_H 22 typename LhsScalar,
int LhsStorageOrder,
bool ConjugateLhs,
23 typename RhsScalar,
int RhsStorageOrder,
bool ConjugateRhs>
28 Index rows, Index cols, Index depth,
29 const LhsScalar* lhs, Index lhsStride,
30 const RhsScalar*
rhs, Index rhsStride,
31 ResScalar* res, Index resStride,
41 ::run(cols,rows,depth,rhs,rhsStride,lhs,lhsStride,res,resStride,alpha,blocking,info);
49 typename LhsScalar,
int LhsStorageOrder,
bool ConjugateLhs,
50 typename RhsScalar,
int RhsStorageOrder,
bool ConjugateRhs>
55 static void run(Index rows, Index cols, Index depth,
56 const LhsScalar* _lhs, Index lhsStride,
57 const RhsScalar* _rhs, Index rhsStride,
58 ResScalar* res, Index resStride,
68 Index kc = blocking.
kc();
69 Index mc = (std::min)(rows,blocking.
mc());
76 #ifdef EIGEN_HAS_OPENMP 80 Index tid = omp_get_thread_num();
81 Index threads = omp_get_num_threads();
83 std::size_t sizeA = kc*mc;
84 std::size_t sizeW = kc*Traits::WorkSpaceFactor;
88 RhsScalar* blockB = blocking.
blockB();
92 for(Index k=0; k<depth; k+=kc)
94 const Index actual_kc = (std::min)(k+kc,depth)-k;
98 pack_lhs(blockA, &lhs(0,k), lhsStride, actual_kc, mc);
106 while(info[tid].users!=0) {}
107 info[tid].users += threads;
109 pack_rhs(blockB+info[tid].rhs_start*actual_kc, &
rhs(k,info[tid].rhs_start), rhsStride, actual_kc, info[tid].rhs_length);
115 for(Index shift=0; shift<threads; ++shift)
117 Index j = (tid+shift)%threads;
123 while(info[j].sync!=k) {}
125 gebp(res+info[j].rhs_start*resStride, resStride, blockA, blockB+info[j].rhs_start*actual_kc, mc, actual_kc, info[j].rhs_length, alpha, -1,-1,0,0, w);
129 for(Index i=mc; i<rows; i+=mc)
131 const Index actual_mc = (std::min)(i+mc,rows)-i;
134 pack_lhs(blockA, &lhs(i,k), lhsStride, actual_kc, actual_mc);
137 gebp(res+i, resStride, blockA, blockB, actual_mc, actual_kc, cols, alpha, -1,-1,0,0, w);
142 for(Index j=0; j<threads; ++j)
148 #endif // EIGEN_HAS_OPENMP 153 std::size_t sizeA = kc*mc;
154 std::size_t sizeB = kc*cols;
155 std::size_t sizeW = kc*Traits::WorkSpaceFactor;
163 for(Index k2=0; k2<depth; k2+=kc)
165 const Index actual_kc = (std::min)(k2+kc,depth)-k2;
171 pack_rhs(blockB, &
rhs(k2,0), rhsStride, actual_kc, cols);
175 for(Index i2=0; i2<rows; i2+=mc)
177 const Index actual_mc = (std::min)(i2+mc,rows)-i2;
182 pack_lhs(blockA, &lhs(i2,k2), lhsStride, actual_kc, actual_mc);
185 gebp(res+i2, resStride, blockA, blockB, actual_mc, actual_kc, cols, alpha, -1, -1, 0, 0, blockW);
198 template<
typename Lhs,
typename Rhs>
200 :
traits<ProductBase<GeneralProduct<Lhs,Rhs,GemmProduct>, Lhs, Rhs> >
203 template<
typename Scalar,
typename Index,
typename Gemm,
typename Lhs,
typename Rhs,
typename Dest,
typename BlockingType>
207 BlockingType& blocking)
208 : m_lhs(lhs), m_rhs(rhs), m_dest(dest), m_actualAlpha(actualAlpha), m_blocking(blocking)
213 m_blocking.allocateB();
221 Gemm::run(rows, cols, m_lhs.cols(),
222 &m_lhs.coeffRef(row,0), m_lhs.outerStride(),
223 &m_rhs.coeffRef(0,
col), m_rhs.outerStride(),
224 (Scalar*)&(m_dest.coeffRef(row,
col)), m_dest.outerStride(),
225 m_actualAlpha, m_blocking, info);
236 template<
int StorageOrder,
typename LhsScalar,
typename RhsScalar,
int MaxRows,
int MaxCols,
int MaxDepth,
int KcFactor=1,
239 template<
typename _LhsScalar,
typename _RhsScalar>
257 : m_blockA(0), m_blockB(0), m_blockW(0), m_mc(0), m_nc(0), m_kc(0)
264 inline LhsScalar*
blockA() {
return m_blockA; }
265 inline RhsScalar*
blockB() {
return m_blockB; }
266 inline RhsScalar*
blockW() {
return m_blockW; }
269 template<
int StorageOrder,
typename _LhsScalar,
typename _RhsScalar,
int MaxRows,
int MaxCols,
int MaxDepth,
int KcFactor>
272 typename conditional<StorageOrder==RowMajor,_RhsScalar,_LhsScalar>::type,
273 typename conditional<StorageOrder==RowMajor,_LhsScalar,_RhsScalar>::type>
284 SizeA = ActualRows * MaxDepth,
285 SizeB = ActualCols * MaxDepth,
286 SizeW = MaxDepth * Traits::WorkSpaceFactor
297 this->m_mc = ActualRows;
298 this->m_nc = ActualCols;
299 this->m_kc = MaxDepth;
300 this->m_blockA = m_staticA;
301 this->m_blockB = m_staticB;
302 this->m_blockW = m_staticW;
311 template<
int StorageOrder,
typename _LhsScalar,
typename _RhsScalar,
int MaxRows,
int MaxCols,
int MaxDepth,
int KcFactor>
314 typename conditional<StorageOrder==RowMajor,_RhsScalar,_LhsScalar>::type,
315 typename conditional<StorageOrder==RowMajor,_LhsScalar,_RhsScalar>::type>
336 computeProductBlockingSizes<LhsScalar,RhsScalar,KcFactor>(this->m_kc, this->m_mc, this->m_nc);
337 m_sizeA = this->m_mc * this->m_kc;
338 m_sizeB = this->m_kc * this->m_nc;
339 m_sizeW = this->m_kc*Traits::WorkSpaceFactor;
344 if(this->m_blockA==0)
345 this->m_blockA = aligned_new<LhsScalar>(m_sizeA);
350 if(this->m_blockB==0)
351 this->m_blockB = aligned_new<RhsScalar>(m_sizeB);
356 if(this->m_blockW==0)
357 this->m_blockW = aligned_new<RhsScalar>(m_sizeW);
377 template<
typename Lhs,
typename Rhs>
379 :
public ProductBase<GeneralProduct<Lhs,Rhs,GemmProduct>, Lhs, Rhs>
399 eigen_assert(dst.rows()==m_lhs.rows() && dst.cols()==m_rhs.cols());
404 Scalar actualAlpha = alpha * LhsBlasTraits::extractScalarFactor(m_lhs)
405 * RhsBlasTraits::extractScalarFactor(m_rhs);
408 Dest::MaxRowsAtCompileTime,Dest::MaxColsAtCompileTime,MaxDepthAtCompileTime> BlockingType;
419 BlockingType blocking(dst.rows(), dst.cols(), lhs.cols());
421 internal::parallelize_gemm<(Dest::MaxRowsAtCompileTime>32 || Dest::MaxRowsAtCompileTime==
Dynamic)>(GemmFunctor(lhs, rhs, dst, actualAlpha, blocking), this->rows(), this->cols(), Dest::Flags&
RowMajorBit);
427 #endif // EIGEN_GENERAL_MATRIX_MATRIX_H scalar_product_traits< LhsScalar, RhsScalar >::ReturnType ResScalar
Expression of the product of two general matrices or vectors.
gebp_traits< LhsScalar, RhsScalar > Traits
internal::remove_all< ActualRhsType >::type _ActualRhsType
#define EIGEN_PRODUCT_PUBLIC_INTERFACE(Derived)
gebp_traits< LhsScalar, RhsScalar > Traits
static void run(Index rows, Index cols, Index depth, const LhsScalar *_lhs, Index lhsStride, const RhsScalar *_rhs, Index rhsStride, ResScalar *res, Index resStride, ResScalar alpha, level3_blocking< LhsScalar, RhsScalar > &blocking, GemmParallelInfo< Index > *info=0)
#define EIGEN_STRONG_INLINE
internal::traits< Derived >::Scalar Scalar
static EIGEN_STRONG_INLINE void run(Index rows, Index cols, Index depth, const LhsScalar *lhs, Index lhsStride, const RhsScalar *rhs, Index rhsStride, ResScalar *res, Index resStride, ResScalar alpha, level3_blocking< RhsScalar, LhsScalar > &blocking, GemmParallelInfo< Index > *info=0)
#define ei_declare_aligned_stack_constructed_variable(TYPE, NAME, SIZE, BUFFER)
internal::traits< Derived >::Index Index
The type of indices.
Expression of the transpose of a matrix.
#define EIGEN_UNUSED_VARIABLE(var)
conditional< Transpose, _LhsScalar, _RhsScalar >::type RhsScalar
iterative scaling algorithm to equilibrate rows and column norms in matrices
#define eigen_internal_assert(x)
#define EIGEN_SIZE_MIN_PREFER_FIXED(a, b)
const unsigned int RowMajorBit
gemm_functor(const Lhs &lhs, const Rhs &rhs, Dest &dest, const Scalar &actualAlpha, BlockingType &blocking)
gemm_blocking_space(DenseIndex rows, DenseIndex cols, DenseIndex depth)
internal::remove_all< ActualLhsType >::type _ActualLhsType
scalar_product_traits< LhsScalar, RhsScalar >::ReturnType ResScalar
conditional< Transpose, _RhsScalar, _LhsScalar >::type LhsScalar
gemm_blocking_space(DenseIndex, DenseIndex, DenseIndex)
conditional< Transpose, _RhsScalar, _LhsScalar >::type LhsScalar
conditional< Transpose, _LhsScalar, _RhsScalar >::type RhsScalar
void scaleAndAddTo(Dest &dst, const Scalar &alpha) const
void rhs(const real_t *x, real_t *f)
void aligned_delete(T *ptr, size_t size)
EIGEN_DEFAULT_DENSE_INDEX_TYPE DenseIndex
BlockingType & m_blocking
void initParallelSession() const
#define EIGEN_CHECK_BINARY_COMPATIBILIY(BINOP, LHS, RHS)