summaryrefslogtreecommitdiffstats
path: root/src/armadillo/include/armadillo_bits/glue_powext_meat.hpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/armadillo/include/armadillo_bits/glue_powext_meat.hpp')
-rw-r--r--src/armadillo/include/armadillo_bits/glue_powext_meat.hpp674
1 files changed, 674 insertions, 0 deletions
diff --git a/src/armadillo/include/armadillo_bits/glue_powext_meat.hpp b/src/armadillo/include/armadillo_bits/glue_powext_meat.hpp
new file mode 100644
index 0000000..700a2cf
--- /dev/null
+++ b/src/armadillo/include/armadillo_bits/glue_powext_meat.hpp
@@ -0,0 +1,674 @@
+// SPDX-License-Identifier: Apache-2.0
+//
+// Copyright 2008-2016 Conrad Sanderson (http://conradsanderson.id.au)
+// Copyright 2008-2016 National ICT Australia (NICTA)
+//
+// Licensed under the Apache License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+// ------------------------------------------------------------------------
+
+
+
+//! \addtogroup glue_powext
+//! @{
+
+
+template<typename T1, typename T2>
+inline
+void
+glue_powext::apply(Mat<typename T1::elem_type>& out, const Glue<T1, T2, glue_powext>& X)
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename T1::elem_type eT;
+
+ const quasi_unwrap<T1> UA(X.A);
+ const quasi_unwrap<T2> UB(X.B);
+
+ const Mat<eT>& A = UA.M;
+ const Mat<eT>& B = UB.M;
+
+ arma_debug_assert_same_size(A, B, "element-wise pow()");
+
+ const bool UA_bad_alias = UA.is_alias(out) && (UA.has_subview); // allow inplace operation
+ const bool UB_bad_alias = UB.is_alias(out);
+
+ if(UA_bad_alias || UB_bad_alias)
+ {
+ Mat<eT> tmp;
+
+ glue_powext::apply(tmp, A, B);
+
+ out.steal_mem(tmp);
+ }
+ else
+ {
+ glue_powext::apply(out, A, B);
+ }
+ }
+
+
+
+template<typename eT>
+inline
+void
+glue_powext::apply(Mat<eT>& out, const Mat<eT>& A, const Mat<eT>& B)
+ {
+ arma_extra_debug_sigprint();
+
+ out.set_size(A.n_rows, A.n_cols);
+
+ const uword N = out.n_elem;
+
+ eT* out_mem = out.memptr();
+ const eT* A_mem = A.memptr();
+ const eT* B_mem = B.memptr();
+
+ if( arma_config::openmp && mp_gate<eT>::eval(N) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = mp_thread_limit::get();
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = eop_aux::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = eop_aux::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ }
+
+
+
+template<typename parent, unsigned int mode, typename T2>
+inline
+Mat<typename parent::elem_type>
+glue_powext::apply
+ (
+ const subview_each1<parent,mode>& X,
+ const Base<typename parent::elem_type,T2>& Y
+ )
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename parent::elem_type eT;
+
+ const parent& A = X.P;
+
+ const uword A_n_rows = A.n_rows;
+ const uword A_n_cols = A.n_cols;
+
+ Mat<eT> out(A_n_rows, A_n_cols, arma_nozeros_indicator());
+
+ const quasi_unwrap<T2> tmp(Y.get_ref());
+ const Mat<eT>& B = tmp.M;
+
+ X.check_size(B);
+
+ const eT* B_mem = B.memptr();
+
+ if(mode == 0) // each column
+ {
+ if( arma_config::openmp && mp_gate<eT>::eval(A.n_elem) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = int( (std::min)(uword(mp_thread_limit::get()), A_n_cols) );
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = eop_aux::pow(A_mem[row], B_mem[row]);
+ }
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = eop_aux::pow(A_mem[row], B_mem[row]);
+ }
+ }
+ }
+ }
+
+ if(mode == 1) // each row
+ {
+ if( arma_config::openmp && mp_gate<eT>::eval(A.n_elem) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = int( (std::min)(uword(mp_thread_limit::get()), A_n_cols) );
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ const eT B_val = B_mem[i];
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = eop_aux::pow(A_mem[row], B_val);
+ }
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ const eT B_val = B_mem[i];
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = eop_aux::pow(A_mem[row], B_val);
+ }
+ }
+ }
+ }
+
+ return out;
+ }
+
+
+
+template<typename T1, typename T2>
+inline
+void
+glue_powext::apply(Cube<typename T1::elem_type>& out, const GlueCube<T1, T2, glue_powext>& X)
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename T1::elem_type eT;
+
+ const unwrap_cube<T1> UA(X.A);
+ const unwrap_cube<T2> UB(X.B);
+
+ const Cube<eT>& A = UA.M;
+ const Cube<eT>& B = UB.M;
+
+ arma_debug_assert_same_size(A, B, "element-wise pow()");
+
+ if(UB.is_alias(out))
+ {
+ Cube<eT> tmp;
+
+ glue_powext::apply(tmp, A, B);
+
+ out.steal_mem(tmp);
+ }
+ else
+ {
+ glue_powext::apply(out, A, B);
+ }
+ }
+
+
+
+template<typename eT>
+inline
+void
+glue_powext::apply(Cube<eT>& out, const Cube<eT>& A, const Cube<eT>& B)
+ {
+ arma_extra_debug_sigprint();
+
+ out.set_size(A.n_rows, A.n_cols, A.n_slices);
+
+ const uword N = out.n_elem;
+
+ eT* out_mem = out.memptr();
+ const eT* A_mem = A.memptr();
+ const eT* B_mem = B.memptr();
+
+ if( arma_config::openmp && mp_gate<eT>::eval(N) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = mp_thread_limit::get();
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = eop_aux::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = eop_aux::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ }
+
+
+
+template<typename eT, typename T2>
+inline
+Cube<eT>
+glue_powext::apply
+ (
+ const subview_cube_each1<eT>& X,
+ const Base<eT,T2>& Y
+ )
+ {
+ arma_extra_debug_sigprint();
+
+ const Cube<eT>& A = X.P;
+
+ const uword A_n_rows = A.n_rows;
+ const uword A_n_cols = A.n_cols;
+ const uword A_n_slices = A.n_slices;
+
+ Cube<eT> out(A_n_rows, A_n_cols, A_n_slices, arma_nozeros_indicator());
+
+ const quasi_unwrap<T2> tmp(Y.get_ref());
+ const Mat<eT>& B = tmp.M;
+
+ X.check_size(B);
+
+ const eT* B_mem = B.memptr();
+ const uword B_n_elem = B.n_elem;
+
+ if( arma_config::openmp && mp_gate<eT>::eval(A.n_elem) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = int( (std::min)(uword(mp_thread_limit::get()), A_n_slices) );
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword s=0; s < A_n_slices; ++s)
+ {
+ const eT* A_slice_mem = A.slice_memptr(s);
+ eT* out_slice_mem = out.slice_memptr(s);
+
+ for(uword i=0; i < B_n_elem; ++i)
+ {
+ out_slice_mem[i] = eop_aux::pow(A_slice_mem[i], B_mem[i]);
+ }
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword s=0; s < A_n_slices; ++s)
+ {
+ const eT* A_slice_mem = A.slice_memptr(s);
+ eT* out_slice_mem = out.slice_memptr(s);
+
+ for(uword i=0; i < B_n_elem; ++i)
+ {
+ out_slice_mem[i] = eop_aux::pow(A_slice_mem[i], B_mem[i]);
+ }
+ }
+ }
+
+ return out;
+ }
+
+
+
+//
+
+
+
+template<typename T1, typename T2>
+inline
+void
+glue_powext_cx::apply(Mat<typename T1::elem_type>& out, const mtGlue<typename T1::elem_type, T1, T2, glue_powext_cx>& X)
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename T1::elem_type eT;
+ typedef typename T1::pod_type T;
+
+ const quasi_unwrap<T1> UA(X.A);
+ const quasi_unwrap<T2> UB(X.B);
+
+ const Mat<eT>& A = UA.M;
+ const Mat< T>& B = UB.M;
+
+ arma_debug_assert_same_size(A, B, "element-wise pow()");
+
+ if(UA.is_alias(out) && (UA.has_subview))
+ {
+ Mat<eT> tmp;
+
+ glue_powext_cx::apply(tmp, A, B);
+
+ out.steal_mem(tmp);
+ }
+ else
+ {
+ glue_powext_cx::apply(out, A, B);
+ }
+ }
+
+
+
+template<typename T>
+inline
+void
+glue_powext_cx::apply(Mat< std::complex<T> >& out, const Mat< std::complex<T> >& A, const Mat<T>& B)
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename std::complex<T> eT;
+
+ out.set_size(A.n_rows, A.n_cols);
+
+ const uword N = out.n_elem;
+
+ eT* out_mem = out.memptr();
+ const eT* A_mem = A.memptr();
+ const T* B_mem = B.memptr();
+
+ if( arma_config::openmp && mp_gate<eT>::eval(N) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = mp_thread_limit::get();
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = std::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = std::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ }
+
+
+
+template<typename parent, unsigned int mode, typename T2>
+inline
+Mat<typename parent::elem_type>
+glue_powext_cx::apply
+ (
+ const subview_each1<parent,mode>& X,
+ const Base<typename T2::elem_type,T2>& Y
+ )
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename parent::elem_type eT;
+ typedef typename parent::pod_type T;
+
+ const parent& A = X.P;
+
+ const uword A_n_rows = A.n_rows;
+ const uword A_n_cols = A.n_cols;
+
+ Mat<eT> out(A_n_rows, A_n_cols, arma_nozeros_indicator());
+
+ const quasi_unwrap<T2> tmp(Y.get_ref());
+ const Mat<T>& B = tmp.M;
+
+ X.check_size(B);
+
+ const T* B_mem = B.memptr();
+
+ if(mode == 0) // each column
+ {
+ if( arma_config::openmp && mp_gate<eT>::eval(A.n_elem) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = int( (std::min)(uword(mp_thread_limit::get()), A_n_cols) );
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = std::pow(A_mem[row], B_mem[row]);
+ }
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = std::pow(A_mem[row], B_mem[row]);
+ }
+ }
+ }
+ }
+
+ if(mode == 1) // each row
+ {
+ if( arma_config::openmp && mp_gate<eT>::eval(A.n_elem) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = int( (std::min)(uword(mp_thread_limit::get()), A_n_cols) );
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ const eT B_val = B_mem[i];
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = std::pow(A_mem[row], B_val);
+ }
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i < A_n_cols; ++i)
+ {
+ const eT* A_mem = A.colptr(i);
+ eT* out_mem = out.colptr(i);
+
+ const eT B_val = B_mem[i];
+
+ for(uword row=0; row < A_n_rows; ++row)
+ {
+ out_mem[row] = std::pow(A_mem[row], B_val);
+ }
+ }
+ }
+ }
+
+ return out;
+ }
+
+
+
+template<typename T1, typename T2>
+inline
+void
+glue_powext_cx::apply(Cube<typename T1::elem_type>& out, const mtGlueCube<typename T1::elem_type,T1,T2,glue_powext_cx>& X)
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename T1::elem_type eT;
+
+ typedef typename get_pod_type<eT>::result T;
+
+ const unwrap_cube<T1> UA(X.A);
+ const unwrap_cube<T2> UB(X.B);
+
+ const Cube<eT>& A = UA.M;
+ const Cube< T>& B = UB.M;
+
+ arma_debug_assert_same_size(A, B, "element-wise pow()");
+
+ glue_powext_cx::apply(out, A, B);
+ }
+
+
+
+template<typename T>
+inline
+void
+glue_powext_cx::apply(Cube< std::complex<T> >& out, const Cube< std::complex<T> >& A, const Cube<T>& B)
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename std::complex<T> eT;
+
+ out.set_size(A.n_rows, A.n_cols, A.n_slices);
+
+ const uword N = out.n_elem;
+
+ eT* out_mem = out.memptr();
+ const eT* A_mem = A.memptr();
+ const T* B_mem = B.memptr();
+
+ if( arma_config::openmp && mp_gate<eT>::eval(N) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = mp_thread_limit::get();
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = std::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword i=0; i<N; ++i)
+ {
+ out_mem[i] = std::pow(A_mem[i], B_mem[i]);
+ }
+ }
+ }
+
+
+
+template<typename T, typename T2>
+inline
+Cube< std::complex<T> >
+glue_powext_cx::apply
+ (
+ const subview_cube_each1< std::complex<T> >& X,
+ const Base<T,T2>& Y
+ )
+ {
+ arma_extra_debug_sigprint();
+
+ typedef typename std::complex<T> eT;
+
+ const Cube<eT>& A = X.P;
+
+ const uword A_n_rows = A.n_rows;
+ const uword A_n_cols = A.n_cols;
+ const uword A_n_slices = A.n_slices;
+
+ Cube<eT> out(A_n_rows, A_n_cols, A_n_slices, arma_nozeros_indicator());
+
+ const quasi_unwrap<T2> tmp(Y.get_ref());
+ const Mat<T>& B = tmp.M;
+
+ X.check_size(B);
+
+ const T* B_mem = B.memptr();
+ const uword B_n_elem = B.n_elem;
+
+ if( arma_config::openmp && mp_gate<eT>::eval(A.n_elem) )
+ {
+ #if defined(ARMA_USE_OPENMP)
+ {
+ const int n_threads = int( (std::min)(uword(mp_thread_limit::get()), A_n_slices) );
+
+ #pragma omp parallel for schedule(static) num_threads(n_threads)
+ for(uword s=0; s < A_n_slices; ++s)
+ {
+ const eT* A_slice_mem = A.slice_memptr(s);
+ eT* out_slice_mem = out.slice_memptr(s);
+
+ for(uword i=0; i < B_n_elem; ++i)
+ {
+ out_slice_mem[i] = std::pow(A_slice_mem[i], B_mem[i]);
+ }
+ }
+ }
+ #endif
+ }
+ else
+ {
+ for(uword s=0; s < A_n_slices; ++s)
+ {
+ const eT* A_slice_mem = A.slice_memptr(s);
+ eT* out_slice_mem = out.slice_memptr(s);
+
+ for(uword i=0; i < B_n_elem; ++i)
+ {
+ out_slice_mem[i] = std::pow(A_slice_mem[i], B_mem[i]);
+ }
+ }
+ }
+
+ return out;
+ }
+
+
+
+//! @}