12 #ifndef EIGEN_ASSIGN_H
13 #define EIGEN_ASSIGN_H
23 template <
typename Derived,
typename OtherDerived>
30 SrcIsAligned = OtherDerived::Flags &
AlignedBit,
31 JointAlignment = bool(DstIsAligned) && bool(SrcIsAligned) ?
Aligned :
Unaligned
36 InnerSize = int(Derived::IsVectorAtCompileTime) ? int(Derived::SizeAtCompileTime)
37 : int(Derived::Flags)&
RowMajorBit ? int(Derived::ColsAtCompileTime)
38 : int(Derived::RowsAtCompileTime),
39 InnerMaxSize = int(Derived::IsVectorAtCompileTime) ? int(Derived::MaxSizeAtCompileTime)
40 : int(Derived::Flags)&
RowMajorBit ? int(Derived::MaxColsAtCompileTime)
41 : int(Derived::MaxRowsAtCompileTime),
42 MaxSizeAtCompileTime = Derived::SizeAtCompileTime,
43 PacketSize = packet_traits<typename Derived::Scalar>::size
47 StorageOrdersAgree = (int(Derived::IsRowMajor) == int(OtherDerived::IsRowMajor)),
48 MightVectorize = StorageOrdersAgree
50 MayInnerVectorize = MightVectorize &&
int(InnerSize)!=
Dynamic && int(InnerSize)%int(PacketSize)==0
51 && int(DstIsAligned) && int(SrcIsAligned),
52 MayLinearize = StorageOrdersAgree && (int(Derived::Flags) & int(OtherDerived::Flags) &
LinearAccessBit),
53 MayLinearVectorize = MightVectorize && MayLinearize && DstHasDirectAccess
54 && (DstIsAligned || MaxSizeAtCompileTime ==
Dynamic),
57 MaySliceVectorize = MightVectorize && DstHasDirectAccess
58 && (int(InnerMaxSize)==
Dynamic || int(InnerMaxSize)>=3*PacketSize)
66 Traversal = int(MayInnerVectorize) ? int(InnerVectorizedTraversal)
67 : int(MayLinearVectorize) ? int(LinearVectorizedTraversal)
68 : int(MaySliceVectorize) ? int(SliceVectorizedTraversal)
69 : int(MayLinearize) ? int(LinearTraversal)
70 : int(DefaultTraversal),
71 Vectorized = int(Traversal) == InnerVectorizedTraversal
72 || int(Traversal) == LinearVectorizedTraversal
73 || int(Traversal) == SliceVectorizedTraversal
78 UnrollingLimit = EIGEN_UNROLLING_LIMIT * (Vectorized ? int(PacketSize) : 1),
79 MayUnrollCompletely = int(Derived::SizeAtCompileTime) !=
Dynamic
80 && int(OtherDerived::CoeffReadCost) !=
Dynamic
81 && int(Derived::SizeAtCompileTime) * int(OtherDerived::CoeffReadCost) <= int(UnrollingLimit),
82 MayUnrollInner = int(InnerSize) !=
Dynamic
83 && int(OtherDerived::CoeffReadCost) !=
Dynamic
84 && int(InnerSize) * int(OtherDerived::CoeffReadCost) <= int(UnrollingLimit)
89 Unrolling = (int(Traversal) == int(InnerVectorizedTraversal) || int(Traversal) == int(DefaultTraversal))
91 int(MayUnrollCompletely) ? int(CompleteUnrolling)
92 : int(MayUnrollInner) ? int(InnerUnrolling)
95 : int(Traversal) == int(LinearVectorizedTraversal)
96 ? ( bool(MayUnrollCompletely) && bool(DstIsAligned) ? int(CompleteUnrolling) : int(NoUnrolling) )
97 : int(Traversal) == int(LinearTraversal)
98 ? ( bool(MayUnrollCompletely) ? int(CompleteUnrolling) : int(NoUnrolling) )
102 #ifdef EIGEN_DEBUG_ASSIGN
105 EIGEN_DEBUG_VAR(DstIsAligned)
106 EIGEN_DEBUG_VAR(SrcIsAligned)
107 EIGEN_DEBUG_VAR(JointAlignment)
108 EIGEN_DEBUG_VAR(InnerSize)
109 EIGEN_DEBUG_VAR(InnerMaxSize)
110 EIGEN_DEBUG_VAR(PacketSize)
111 EIGEN_DEBUG_VAR(StorageOrdersAgree)
112 EIGEN_DEBUG_VAR(MightVectorize)
113 EIGEN_DEBUG_VAR(MayLinearize)
114 EIGEN_DEBUG_VAR(MayInnerVectorize)
115 EIGEN_DEBUG_VAR(MayLinearVectorize)
116 EIGEN_DEBUG_VAR(MaySliceVectorize)
117 EIGEN_DEBUG_VAR(Traversal)
118 EIGEN_DEBUG_VAR(UnrollingLimit)
119 EIGEN_DEBUG_VAR(MayUnrollCompletely)
120 EIGEN_DEBUG_VAR(MayUnrollInner)
121 EIGEN_DEBUG_VAR(Unrolling)
134 template<
typename Derived1,
typename Derived2,
int Index,
int Stop>
135 struct assign_DefaultTraversal_CompleteUnrolling
138 outer = Index / Derived1::InnerSizeAtCompileTime,
139 inner = Index % Derived1::InnerSizeAtCompileTime
142 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
144 dst.copyCoeffByOuterInner(outer, inner, src);
145 assign_DefaultTraversal_CompleteUnrolling<Derived1, Derived2, Index+1, Stop>::run(dst, src);
149 template<
typename Derived1,
typename Derived2,
int Stop>
150 struct assign_DefaultTraversal_CompleteUnrolling<Derived1, Derived2, Stop, Stop>
152 static EIGEN_STRONG_INLINE
void run(Derived1 &,
const Derived2 &) {}
155 template<
typename Derived1,
typename Derived2,
int Index,
int Stop>
156 struct assign_DefaultTraversal_InnerUnrolling
158 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src,
int outer)
160 dst.copyCoeffByOuterInner(outer, Index, src);
161 assign_DefaultTraversal_InnerUnrolling<Derived1, Derived2, Index+1, Stop>::run(dst, src, outer);
165 template<
typename Derived1,
typename Derived2,
int Stop>
166 struct assign_DefaultTraversal_InnerUnrolling<Derived1, Derived2, Stop, Stop>
168 static EIGEN_STRONG_INLINE
void run(Derived1 &,
const Derived2 &,
int) {}
175 template<
typename Derived1,
typename Derived2,
int Index,
int Stop>
176 struct assign_LinearTraversal_CompleteUnrolling
178 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
180 dst.copyCoeff(Index, src);
181 assign_LinearTraversal_CompleteUnrolling<Derived1, Derived2, Index+1, Stop>::run(dst, src);
185 template<
typename Derived1,
typename Derived2,
int Stop>
186 struct assign_LinearTraversal_CompleteUnrolling<Derived1, Derived2, Stop, Stop>
188 static EIGEN_STRONG_INLINE
void run(Derived1 &,
const Derived2 &) {}
195 template<
typename Derived1,
typename Derived2,
int Index,
int Stop>
196 struct assign_innervec_CompleteUnrolling
199 outer = Index / Derived1::InnerSizeAtCompileTime,
200 inner = Index % Derived1::InnerSizeAtCompileTime,
201 JointAlignment = assign_traits<Derived1,Derived2>::JointAlignment
204 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
206 dst.template copyPacketByOuterInner<Derived2, Aligned, JointAlignment>(outer, inner, src);
207 assign_innervec_CompleteUnrolling<Derived1, Derived2,
208 Index+packet_traits<typename Derived1::Scalar>::size, Stop>::run(dst, src);
212 template<
typename Derived1,
typename Derived2,
int Stop>
213 struct assign_innervec_CompleteUnrolling<Derived1, Derived2, Stop, Stop>
215 static EIGEN_STRONG_INLINE
void run(Derived1 &,
const Derived2 &) {}
218 template<
typename Derived1,
typename Derived2,
int Index,
int Stop>
219 struct assign_innervec_InnerUnrolling
221 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src,
int outer)
223 dst.template copyPacketByOuterInner<Derived2, Aligned, Aligned>(outer, Index, src);
224 assign_innervec_InnerUnrolling<Derived1, Derived2,
225 Index+packet_traits<typename Derived1::Scalar>::size, Stop>::run(dst, src, outer);
229 template<
typename Derived1,
typename Derived2,
int Stop>
230 struct assign_innervec_InnerUnrolling<Derived1, Derived2, Stop, Stop>
232 static EIGEN_STRONG_INLINE
void run(Derived1 &,
const Derived2 &,
int) {}
239 template<
typename Derived1,
typename Derived2,
240 int Traversal = assign_traits<Derived1, Derived2>::Traversal,
241 int Unrolling = assign_traits<Derived1, Derived2>::Unrolling,
242 int Version = Specialized>
249 template<
typename Derived1,
typename Derived2,
int Unrolling,
int Version>
250 struct assign_impl<Derived1, Derived2, InvalidTraversal, Unrolling, Version>
252 static inline void run(Derived1 &,
const Derived2 &) { }
255 template<
typename Derived1,
typename Derived2,
int Version>
256 struct assign_impl<Derived1, Derived2, DefaultTraversal, NoUnrolling, Version>
258 typedef typename Derived1::Index Index;
259 static inline void run(Derived1 &dst,
const Derived2 &src)
261 const Index innerSize = dst.innerSize();
262 const Index outerSize = dst.outerSize();
263 for(Index outer = 0; outer < outerSize; ++outer)
264 for(Index inner = 0; inner < innerSize; ++inner)
265 dst.copyCoeffByOuterInner(outer, inner, src);
269 template<
typename Derived1,
typename Derived2,
int Version>
270 struct assign_impl<Derived1, Derived2, DefaultTraversal, CompleteUnrolling, Version>
272 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
274 assign_DefaultTraversal_CompleteUnrolling<Derived1, Derived2, 0, Derived1::SizeAtCompileTime>
279 template<
typename Derived1,
typename Derived2,
int Version>
280 struct assign_impl<Derived1, Derived2, DefaultTraversal, InnerUnrolling, Version>
282 typedef typename Derived1::Index Index;
283 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
285 const Index outerSize = dst.outerSize();
286 for(Index outer = 0; outer < outerSize; ++outer)
287 assign_DefaultTraversal_InnerUnrolling<Derived1, Derived2, 0, Derived1::InnerSizeAtCompileTime>
288 ::run(dst, src, outer);
296 template<
typename Derived1,
typename Derived2,
int Version>
297 struct assign_impl<Derived1, Derived2, LinearTraversal, NoUnrolling, Version>
299 typedef typename Derived1::Index Index;
300 static inline void run(Derived1 &dst,
const Derived2 &src)
302 const Index size = dst.size();
303 for(Index i = 0; i < size; ++i)
304 dst.copyCoeff(i, src);
308 template<
typename Derived1,
typename Derived2,
int Version>
309 struct assign_impl<Derived1, Derived2, LinearTraversal, CompleteUnrolling, Version>
311 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
313 assign_LinearTraversal_CompleteUnrolling<Derived1, Derived2, 0, Derived1::SizeAtCompileTime>
322 template<
typename Derived1,
typename Derived2,
int Version>
323 struct assign_impl<Derived1, Derived2, InnerVectorizedTraversal, NoUnrolling, Version>
325 typedef typename Derived1::Index Index;
326 static inline void run(Derived1 &dst,
const Derived2 &src)
328 const Index innerSize = dst.innerSize();
329 const Index outerSize = dst.outerSize();
330 const Index packetSize = packet_traits<typename Derived1::Scalar>::size;
331 for(Index outer = 0; outer < outerSize; ++outer)
332 for(Index inner = 0; inner < innerSize; inner+=packetSize)
333 dst.template copyPacketByOuterInner<Derived2, Aligned, Aligned>(outer, inner, src);
337 template<
typename Derived1,
typename Derived2,
int Version>
338 struct assign_impl<Derived1, Derived2, InnerVectorizedTraversal, CompleteUnrolling, Version>
340 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
342 assign_innervec_CompleteUnrolling<Derived1, Derived2, 0, Derived1::SizeAtCompileTime>
347 template<
typename Derived1,
typename Derived2,
int Version>
348 struct assign_impl<Derived1, Derived2, InnerVectorizedTraversal, InnerUnrolling, Version>
350 typedef typename Derived1::Index Index;
351 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
353 const Index outerSize = dst.outerSize();
354 for(Index outer = 0; outer < outerSize; ++outer)
355 assign_innervec_InnerUnrolling<Derived1, Derived2, 0, Derived1::InnerSizeAtCompileTime>
356 ::run(dst, src, outer);
364 template <
bool IsAligned = false>
365 struct unaligned_assign_impl
367 template <
typename Derived,
typename OtherDerived>
368 static EIGEN_STRONG_INLINE
void run(
const Derived&, OtherDerived&,
typename Derived::Index,
typename Derived::Index) {}
372 struct unaligned_assign_impl<false>
377 template <
typename Derived,
typename OtherDerived>
378 static EIGEN_DONT_INLINE
void run(
const Derived& src, OtherDerived& dst,
typename Derived::Index start,
typename Derived::Index end)
380 template <
typename Derived,
typename OtherDerived>
381 static EIGEN_STRONG_INLINE
void run(
const Derived& src, OtherDerived& dst,
typename Derived::Index start,
typename Derived::Index end)
384 for (
typename Derived::Index index = start; index < end; ++index)
385 dst.copyCoeff(index, src);
389 template<
typename Derived1,
typename Derived2,
int Version>
390 struct assign_impl<Derived1, Derived2, LinearVectorizedTraversal, NoUnrolling, Version>
392 typedef typename Derived1::Index Index;
393 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
395 const Index size = dst.size();
396 typedef packet_traits<typename Derived1::Scalar> PacketTraits;
398 packetSize = PacketTraits::size,
399 dstAlignment = PacketTraits::AlignedOnScalar ?
Aligned : int(assign_traits<Derived1,Derived2>::DstIsAligned) ,
400 srcAlignment = assign_traits<Derived1,Derived2>::JointAlignment
402 const Index alignedStart = assign_traits<Derived1,Derived2>::DstIsAligned ? 0
403 : internal::first_aligned(&dst.coeffRef(0), size);
404 const Index alignedEnd = alignedStart + ((size-alignedStart)/packetSize)*packetSize;
406 unaligned_assign_impl<assign_traits<Derived1,Derived2>::DstIsAligned!=0>::run(src,dst,0,alignedStart);
408 for(Index index = alignedStart; index < alignedEnd; index += packetSize)
410 dst.template copyPacket<Derived2, dstAlignment, srcAlignment>(index, src);
413 unaligned_assign_impl<>::run(src,dst,alignedEnd,size);
417 template<
typename Derived1,
typename Derived2,
int Version>
418 struct assign_impl<Derived1, Derived2, LinearVectorizedTraversal, CompleteUnrolling, Version>
420 typedef typename Derived1::Index Index;
421 static EIGEN_STRONG_INLINE
void run(Derived1 &dst,
const Derived2 &src)
423 enum { size = Derived1::SizeAtCompileTime,
424 packetSize = packet_traits<typename Derived1::Scalar>::size,
425 alignedSize = (size/packetSize)*packetSize };
427 assign_innervec_CompleteUnrolling<Derived1, Derived2, 0, alignedSize>::run(dst, src);
428 assign_DefaultTraversal_CompleteUnrolling<Derived1, Derived2, alignedSize, size>::run(dst, src);
436 template<
typename Derived1,
typename Derived2,
int Version>
437 struct assign_impl<Derived1, Derived2, SliceVectorizedTraversal, NoUnrolling, Version>
439 typedef typename Derived1::Index Index;
440 static inline void run(Derived1 &dst,
const Derived2 &src)
442 typedef packet_traits<typename Derived1::Scalar> PacketTraits;
444 packetSize = PacketTraits::size,
445 alignable = PacketTraits::AlignedOnScalar,
446 dstAlignment = alignable ?
Aligned : int(assign_traits<Derived1,Derived2>::DstIsAligned) ,
447 srcAlignment = assign_traits<Derived1,Derived2>::JointAlignment
449 const Index packetAlignedMask = packetSize - 1;
450 const Index innerSize = dst.innerSize();
451 const Index outerSize = dst.outerSize();
452 const Index alignedStep = alignable ? (packetSize - dst.outerStride() % packetSize) & packetAlignedMask : 0;
453 Index alignedStart = ((!alignable) || assign_traits<Derived1,Derived2>::DstIsAligned) ? 0
454 : internal::first_aligned(&dst.coeffRef(0,0), innerSize);
456 for(Index outer = 0; outer < outerSize; ++outer)
458 const Index alignedEnd = alignedStart + ((innerSize-alignedStart) & ~packetAlignedMask);
460 for(Index inner = 0; inner<alignedStart ; ++inner)
461 dst.copyCoeffByOuterInner(outer, inner, src);
464 for(Index inner = alignedStart; inner<alignedEnd; inner+=packetSize)
465 dst.template copyPacketByOuterInner<Derived2, dstAlignment, Unaligned>(outer, inner, src);
468 for(Index inner = alignedEnd; inner<innerSize ; ++inner)
469 dst.copyCoeffByOuterInner(outer, inner, src);
471 alignedStart = std::min<Index>((alignedStart+alignedStep)%packetSize, innerSize);
482 template<
typename Derived>
483 template<
typename OtherDerived>
484 EIGEN_STRONG_INLINE Derived& DenseBase<Derived>
485 ::lazyAssign(
const DenseBase<OtherDerived>& other)
488 SameType = internal::is_same<typename Derived::Scalar,typename OtherDerived::Scalar>::value
491 EIGEN_STATIC_ASSERT_LVALUE(Derived)
492 EIGEN_STATIC_ASSERT_SAME_MATRIX_SIZE(Derived,OtherDerived)
493 EIGEN_STATIC_ASSERT(SameType,YOU_MIXED_DIFFERENT_NUMERIC_TYPES__YOU_NEED_TO_USE_THE_CAST_METHOD_OF_MATRIXBASE_TO_CAST_NUMERIC_TYPES_EXPLICITLY)
495 #ifdef EIGEN_DEBUG_ASSIGN
496 internal::assign_traits<Derived, OtherDerived>::debug();
498 eigen_assert(rows() == other.rows() && cols() == other.cols());
499 internal::assign_impl<Derived, OtherDerived, int(SameType) ? int(internal::assign_traits<Derived, OtherDerived>::Traversal)
500 :
int(InvalidTraversal)>::run(derived(),other.derived());
501 #ifndef EIGEN_NO_DEBUG
502 checkTransposeAliasing(other.derived());
509 template<
typename Derived,
typename OtherDerived,
511 bool NeedToTranspose = Derived::IsVectorAtCompileTime
512 && OtherDerived::IsVectorAtCompileTime
513 && ((
int(Derived::RowsAtCompileTime) == 1 &&
int(OtherDerived::ColsAtCompileTime) == 1)
516 (int(Derived::ColsAtCompileTime) == 1 && int(OtherDerived::RowsAtCompileTime) == 1))
517 && int(Derived::SizeAtCompileTime) != 1>
518 struct assign_selector;
520 template<
typename Derived,
typename OtherDerived>
521 struct assign_selector<Derived,OtherDerived,false,false> {
522 static EIGEN_STRONG_INLINE Derived& run(Derived& dst,
const OtherDerived& other) {
return dst.lazyAssign(other.derived()); }
524 template<
typename Derived,
typename OtherDerived>
525 struct assign_selector<Derived,OtherDerived,true,false> {
526 static EIGEN_STRONG_INLINE Derived& run(Derived& dst,
const OtherDerived& other) {
return dst.lazyAssign(other.eval()); }
528 template<
typename Derived,
typename OtherDerived>
529 struct assign_selector<Derived,OtherDerived,false,true> {
530 static EIGEN_STRONG_INLINE Derived& run(Derived& dst,
const OtherDerived& other) {
return dst.lazyAssign(other.transpose()); }
532 template<
typename Derived,
typename OtherDerived>
533 struct assign_selector<Derived,OtherDerived,true,true> {
534 static EIGEN_STRONG_INLINE Derived& run(Derived& dst,
const OtherDerived& other) {
return dst.lazyAssign(other.transpose().eval()); }
539 template<
typename Derived>
540 template<
typename OtherDerived>
543 return internal::assign_selector<Derived,OtherDerived>::run(derived(), other.derived());
546 template<
typename Derived>
549 return internal::assign_selector<Derived,Derived>::run(derived(), other.derived());
552 template<
typename Derived>
555 return internal::assign_selector<Derived,Derived>::run(derived(), other.derived());
558 template<
typename Derived>
559 template <
typename OtherDerived>
562 return internal::assign_selector<Derived,OtherDerived>::run(derived(), other.derived());
565 template<
typename Derived>
566 template <
typename OtherDerived>
569 other.derived().evalTo(derived());
573 template<
typename Derived>
574 template<
typename OtherDerived>
577 other.evalTo(derived());
583 #endif // EIGEN_ASSIGN_H