46 |
|
* [4] Vardeman & Gezelter, in progress (2009). |
47 |
|
*/ |
48 |
|
|
49 |
< |
#ifndef FORCEDECOMPOSITION_COMMUNICATOR_HPP |
50 |
< |
#define FORCEDECOMPOSITION_COMMUNICATOR_HPP |
49 |
> |
#ifndef PARALLEL_COMMUNICATOR_HPP |
50 |
> |
#define PARALLEL_COMMUNICATOR_HPP |
51 |
|
|
52 |
|
#include <config.h> |
53 |
|
#include <mpi.h> |
57 |
|
|
58 |
|
#ifdef IS_MPI |
59 |
|
|
60 |
< |
enum direction { |
61 |
< |
Row = 0, |
62 |
< |
Column = 1 |
60 |
> |
enum communicatorType { |
61 |
> |
Global = 0, |
62 |
> |
Row = 1, |
63 |
> |
Column = 2 |
64 |
|
}; |
65 |
|
|
66 |
|
template<typename T> |
79 |
|
template<> const MPI::Datatype MPITraits<Mat3x3d>::datatype = MPI_REALTYPE; |
80 |
|
template<> const int MPITraits<Mat3x3d>::dim = 9; |
81 |
|
|
82 |
< |
template<direction D, typename T> |
82 |
> |
template<communicatorType D, typename T> |
83 |
|
class Communicator { |
84 |
|
public: |
85 |
|
|
99 |
|
rowIndex_ = myRank / nColumns; |
100 |
|
columnIndex_ = myRank % nColumns; |
101 |
|
|
102 |
< |
if (D == Row) { |
102 |
> |
switch(D) { |
103 |
> |
case Row : |
104 |
|
myComm = MPI::COMM_WORLD.Split(rowIndex_, 0); |
105 |
< |
} else { |
105 |
> |
break; |
106 |
> |
case Column: |
107 |
|
myComm = MPI::COMM_WORLD.Split(columnIndex_, 0); |
108 |
+ |
break; |
109 |
+ |
case Global: |
110 |
+ |
myComm = MPI::COMM_WORLD.Split(myRank, 0); |
111 |
|
} |
112 |
|
|
113 |
|
int nCommProcs = myComm.Get_size(); |
119 |
|
|
120 |
|
myComm.Allgather(&planSize_, 1, MPI::INT, &counts[0], 1, MPI::INT); |
121 |
|
|
116 |
– |
|
122 |
|
displacements[0] = 0; |
123 |
|
for (int i = 1; i < nCommProcs; i++) { |
124 |
|
displacements[i] = displacements[i-1] + counts[i-1]; |
125 |
< |
size_ += count[i-1]; |
125 |
> |
size_ += counts[i-1]; |
126 |
|
} |
127 |
|
|
128 |
|
size_ = 0; |