From d0e130b200988d4a7cbce57f9bd312a7b5258f95 Mon Sep 17 00:00:00 2001
From: Florent Pruvost <florent.pruvost@inria.fr>
Date: Fri, 21 Nov 2014 14:11:55 +0000
Subject: [PATCH] update texinfo users_guide: caution about compatibility

---
 coreblas/include/CMakeLists.txt       |  3 ++
 docs/texinfo/chapters/installing.texi | 48 +++++++++++++++++++++++++++
 2 files changed, 51 insertions(+)

diff --git a/coreblas/include/CMakeLists.txt b/coreblas/include/CMakeLists.txt
index bc878de26..dae6d9202 100644
--- a/coreblas/include/CMakeLists.txt
+++ b/coreblas/include/CMakeLists.txt
@@ -40,6 +40,9 @@ precisions_rules_py(COREBLAS_HDRS_GENERATED "${ZHDR}"
 # --------------------------
 set(COREBLAS_HDRS
     cblas.h
+    lapacke.h
+    lapacke_config.h
+    lapacke_mangling.h
     coreblas.h
     ${COREBLAS_HDRS_GENERATED}
     )
diff --git a/docs/texinfo/chapters/installing.texi b/docs/texinfo/chapters/installing.texi
index 1df7e8e89..f1989c7e9 100644
--- a/docs/texinfo/chapters/installing.texi
+++ b/docs/texinfo/chapters/installing.texi
@@ -71,6 +71,10 @@ Fast implementations are also available as academic packages, such as ATLAS and
 Goto BLAS. 
 The standard interface to BLAS is the FORTRAN interface.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+the reference BLAS from NETLIB and the Intel MKL 11.1 from Intel distribution 
+2013_sp1.
+
 @node CBLAS
 @subsubsection CBLAS
 
@@ -81,6 +85,10 @@ Netlib provides a reference implementation of CBLAS on top of FORTRAN BLAS
 (Netlib CBLAS). 
 Since GSL is implemented in C, it naturally provides CBLAS.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+the reference CBLAS from NETLIB and the Intel MKL 11.1 from Intel distribution 
+2013_sp1.
+
 @node a LAPACK implementation
 @subsubsection a LAPACK implementation
 
@@ -91,6 +99,10 @@ LAPACK provides routines for solving linear systems of equations, linear least
 square problems, eigenvalue problems and singular value problems. 
 Most commercial and academic BLAS packages also provide some LAPACK routines.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+the reference LAPACK from NETLIB and the Intel MKL 11.1 from Intel distribution 
+2013_sp1.
+
 @node LAPACKE
 @subsubsection LAPACKE
 
@@ -103,6 +115,10 @@ In addition to implementing the C interface, LAPACKE also provides routines
 which automatically handle workspace allocation, making the use of LAPACK much 
 more convenient.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+the reference LAPACKE from NETLIB. 
+A stand-alone version of LAPACKE is required.
+
 @node libtmg
 @subsubsection libtmg
 
@@ -112,6 +128,10 @@ of input matrices for testing and timing of LAPACK.
 The testing and timing suites of LAPACK require libtmg, but not the library 
 itself. Note that the LAPACK library can be built and used without libtmg.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+the reference TMG from NETLIB and the Intel MKL 11.1 from Intel distribution 
+2013_sp1.
+
 @node QUARK
 @subsubsection QUARK
 
@@ -126,6 +146,9 @@ When CHAMELEON is linked with QUARK, it is not possible to exploit neither
 CUDA (for GPUs) nor MPI (distributed-memory environment).
 You can use StarPU to do so.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+the QUARK library from PLASMA release between versions 2.5.0 and 2.6.0.
+
 @node StarPU
 @subsubsection StarPU
 
@@ -145,6 +168,9 @@ tasks on the architecture.
 If StarPU is enabled then QUARK is disabled and conversely.
 Note StarPU is enabled by default.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+StarPU-1.1 releases.
+
 @node hwloc
 @subsubsection hwloc
 
@@ -158,6 +184,9 @@ scheduling.
 @code{hwloc} is available in major distributions and for most OSes and can be  
 downloaded from @uref{http://www.open-mpi.org/software/hwloc}.
 
+@strong{Caution about the compatibility:} hwloc should be compatible with the 
+version of StarPU used.
+
 @node pthread
 @subsubsection pthread
 
@@ -185,6 +214,9 @@ MPI can be enabled only if the runtime system chosen is StarPU (default).
 To use MPI through StarPU, it is necessary to compile StarPU with MPI 
 enabled.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+OpenMPI releases from versions 1.4 to 1.6.
+
 @node Nvidia CUDA Toolkit
 @subsubsection Nvidia CUDA Toolkit
 
@@ -203,6 +235,10 @@ is StarPU (default).
 To use CUDA through StarPU, it is necessary to compile StarPU with CUDA 
 enabled.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+CUDA releases from versions 4 to 6. 
+MAGMA library must be compatible with CUDA.
+
 @node MAGMA
 @subsubsection MAGMA
 
@@ -214,6 +250,15 @@ computations on GPUs.
 To fully benefit from GPUs, the user should enable MAGMA in addition to 
 CUDA/cuBLAS.
 
+@strong{Caution about the compatibility:} CHAMELEON has been mainly tested with 
+MAGMA releases from versions 1.4 to 1.6.
+MAGMA library must be compatible with CUDA.
+MAGMA library should be built with sequential versions of BLAS/LAPACK. 
+We should not get some MAGMA link flags embarking multithreaded 
+BLAS/LAPACK because it could affect permformances (take care about the 
+MAGMA link flag @option{-lmkl_intel_thread} for example that we could heritate 
+from the pkg-config file @file{magma.pc}).
+
 @node FxT
 @subsubsection FxT
 
@@ -226,6 +271,9 @@ FxT can only be used through StarPU and StarPU must be compiled with FxT
 enabled, see how to use this feature here @ref{Use FxT profiling through 
 StarPU}.
 
+@strong{Caution about the compatibility:} FxT should be compatible with the 
+version of StarPU used.
+
 @node Build process of CHAMELEON
 @section Build process of CHAMELEON
 
-- 
GitLab