1 |
< |
/* |
1 |
> |
/* |
2 |
|
* Copyright (c) 2005 The University of Notre Dame. All Rights Reserved. |
3 |
|
* |
4 |
|
* The University of Notre Dame grants you ("Licensee") a |
42 |
|
#include "math/ParallelRandNumGen.hpp" |
43 |
|
#ifdef IS_MPI |
44 |
|
#include <mpi.h> |
45 |
+ |
#endif |
46 |
|
|
47 |
|
namespace oopse { |
48 |
|
|
49 |
+ |
int ParallelRandNumGen::nCreatedRNG_ = 0; |
50 |
|
|
51 |
+ |
ParallelRandNumGen::ParallelRandNumGen(const uint32& oneSeed) { |
52 |
|
|
50 |
– |
int ParallelRandNumGen::nCreatedRNG_ = 0; |
51 |
– |
|
52 |
– |
ParallelRandNumGen::ParallelRandNumGen( const uint32& oneSeed) { |
53 |
– |
|
53 |
|
const int masterNode = 0; |
54 |
< |
int seed = oneSeed; |
54 |
> |
unsigned long seed = oneSeed; |
55 |
> |
|
56 |
> |
#ifdef IS_MPI |
57 |
|
MPI_Bcast(&seed, 1, MPI_UNSIGNED_LONG, masterNode, MPI_COMM_WORLD); |
58 |
+ |
#endif |
59 |
|
|
60 |
|
if (seed != oneSeed) { |
61 |
< |
sprintf(painCave.errMsg, |
62 |
< |
"Using different seed to initialize ParallelRandNumGen.\n"); |
63 |
< |
painCave.isFatal = 1;; |
64 |
< |
simError(); |
61 |
> |
sprintf(painCave.errMsg, |
62 |
> |
"Using different seed to initialize ParallelRandNumGen.\n"); |
63 |
> |
painCave.isFatal = 1;; |
64 |
> |
simError(); |
65 |
|
} |
66 |
|
|
67 |
|
int nProcessors; |
68 |
< |
MPI_Comm_size(MPI_COMM_WORLD, &nProcessors); |
68 |
> |
#ifdef IS_MPI |
69 |
> |
MPI_Comm_size( MPI_COMM_WORLD, &nProcessors); |
70 |
|
MPI_Comm_rank( MPI_COMM_WORLD, &myRank_); |
71 |
< |
//In order to generate independent random number stream, the actual seed used by random |
72 |
< |
//number generator is the seed passed to the constructor plus the number of random number |
73 |
< |
//generators which are already created. |
74 |
< |
int newSeed = oneSeed + nCreatedRNG_; |
71 |
> |
#else |
72 |
> |
nProcessors = 1; |
73 |
> |
myRank_ = 0; |
74 |
> |
#endif |
75 |
> |
//In order to generate independent random number stream, the |
76 |
> |
//actual seed used by random number generator is the seed passed |
77 |
> |
//to the constructor plus the number of random number generators |
78 |
> |
//which are already created. |
79 |
> |
unsigned long newSeed = oneSeed + nCreatedRNG_; |
80 |
|
mtRand_ = new MTRand(newSeed, nProcessors, myRank_); |
81 |
< |
|
81 |
> |
|
82 |
|
++nCreatedRNG_; |
83 |
< |
} |
83 |
> |
} |
84 |
|
|
85 |
< |
ParallelRandNumGen::ParallelRandNumGen() { |
85 |
> |
ParallelRandNumGen::ParallelRandNumGen() { |
86 |
|
|
87 |
|
std::vector<uint32> bigSeed; |
88 |
|
const int masterNode = 0; |
89 |
|
int nProcessors; |
90 |
< |
MPI_Comm_size(MPI_COMM_WORLD, &nProcessors); |
90 |
> |
#ifdef IS_MPI |
91 |
> |
MPI_Comm_size( MPI_COMM_WORLD, &nProcessors); |
92 |
|
MPI_Comm_rank( MPI_COMM_WORLD, &myRank_); |
93 |
+ |
#else |
94 |
+ |
nProcessors = 1; |
95 |
+ |
myRank_ = 0; |
96 |
+ |
#endif |
97 |
|
mtRand_ = new MTRand(nProcessors, myRank_); |
98 |
|
|
99 |
< |
seed(); /** @todo calling virtual function in constructor is not a good design */ |
100 |
< |
} |
99 |
> |
seed(); /** @todo calling virtual function in constructor is |
100 |
> |
not a good design */ |
101 |
> |
} |
102 |
|
|
103 |
|
|
104 |
< |
void ParallelRandNumGen::seed( const uint32 oneSeed ) { |
104 |
> |
void ParallelRandNumGen::seed( const uint32 oneSeed ) { |
105 |
|
|
106 |
|
const int masterNode = 0; |
107 |
< |
int seed = oneSeed; |
107 |
> |
unsigned long seed = oneSeed; |
108 |
> |
#ifdef IS_MPI |
109 |
|
MPI_Bcast(&seed, 1, MPI_UNSIGNED_LONG, masterNode, MPI_COMM_WORLD); |
110 |
< |
|
110 |
> |
#endif |
111 |
|
if (seed != oneSeed) { |
112 |
< |
sprintf(painCave.errMsg, |
113 |
< |
"Using different seed to initialize ParallelRandNumGen.\n"); |
114 |
< |
painCave.isFatal = 1;; |
115 |
< |
simError(); |
112 |
> |
sprintf(painCave.errMsg, |
113 |
> |
"Using different seed to initialize ParallelRandNumGen.\n"); |
114 |
> |
painCave.isFatal = 1;; |
115 |
> |
simError(); |
116 |
|
} |
117 |
< |
|
118 |
< |
int newSeed = oneSeed +nCreatedRNG_; |
117 |
> |
|
118 |
> |
unsigned long newSeed = oneSeed +nCreatedRNG_; |
119 |
|
mtRand_->seed(newSeed); |
120 |
< |
|
120 |
> |
|
121 |
|
++nCreatedRNG_; |
122 |
< |
} |
122 |
> |
} |
123 |
|
|
124 |
< |
void ParallelRandNumGen::seed() { |
124 |
> |
void ParallelRandNumGen::seed() { |
125 |
|
|
126 |
|
std::vector<uint32> bigSeed; |
127 |
|
int size; |
128 |
|
const int masterNode = 0; |
129 |
+ |
#ifdef IS_MPI |
130 |
|
if (worldRank == masterNode) { |
131 |
< |
bigSeed = mtRand_->generateSeeds(); |
116 |
< |
size = bigSeed.size(); |
117 |
< |
MPI_Bcast(&size, 1, MPI_INT, masterNode, MPI_COMM_WORLD); |
118 |
< |
MPI_Bcast(&bigSeed[0], size, MPI_UNSIGNED_LONG, masterNode, MPI_COMM_WORLD); |
131 |
> |
#endif |
132 |
|
|
133 |
+ |
bigSeed = mtRand_->generateSeeds(); |
134 |
+ |
size = bigSeed.size(); |
135 |
+ |
|
136 |
+ |
#ifdef IS_MPI |
137 |
+ |
MPI_Bcast(&size, 1, MPI_INT, masterNode, MPI_COMM_WORLD); |
138 |
+ |
MPI_Bcast(&bigSeed[0], size, MPI_UNSIGNED_LONG, masterNode, MPI_COMM_WORLD); |
139 |
|
}else { |
140 |
< |
MPI_Bcast(&size, 1, MPI_INT, masterNode, MPI_COMM_WORLD); |
141 |
< |
bigSeed.resize(size); |
142 |
< |
MPI_Bcast(&bigSeed[0], size, MPI_UNSIGNED_LONG, masterNode, MPI_COMM_WORLD); |
140 |
> |
MPI_Bcast(&size, 1, MPI_INT, masterNode, MPI_COMM_WORLD); |
141 |
> |
bigSeed.resize(size); |
142 |
> |
MPI_Bcast(&bigSeed[0], size, MPI_UNSIGNED_LONG, masterNode, MPI_COMM_WORLD); |
143 |
|
} |
144 |
+ |
#endif |
145 |
|
|
146 |
|
if (bigSeed.size() == 1) { |
147 |
< |
mtRand_->seed(bigSeed[0]); |
147 |
> |
mtRand_->seed(bigSeed[0]); |
148 |
|
} else { |
149 |
< |
mtRand_->seed(&bigSeed[0], bigSeed.size()); |
149 |
> |
mtRand_->seed(&bigSeed[0], bigSeed.size()); |
150 |
|
} |
151 |
|
|
152 |
|
++nCreatedRNG_; |
153 |
< |
} |
134 |
< |
|
135 |
< |
|
153 |
> |
} |
154 |
|
} |
137 |
– |
|
138 |
– |
#endif |