| 1 | // | 
|---|
| 2 | // memmtmpi.cc | 
|---|
| 3 | // based on memmpi.cc | 
|---|
| 4 | // | 
|---|
| 5 | // Copyright (C) 1996 Limit Point Systems, Inc. | 
|---|
| 6 | // | 
|---|
| 7 | // Author: Curtis Janssen <cljanss@limitpt.com> | 
|---|
| 8 | // Maintainer: LPS | 
|---|
| 9 | // | 
|---|
| 10 | // This file is part of the SC Toolkit. | 
|---|
| 11 | // | 
|---|
| 12 | // The SC Toolkit is free software; you can redistribute it and/or modify | 
|---|
| 13 | // it under the terms of the GNU Library General Public License as published by | 
|---|
| 14 | // the Free Software Foundation; either version 2, or (at your option) | 
|---|
| 15 | // any later version. | 
|---|
| 16 | // | 
|---|
| 17 | // The SC Toolkit is distributed in the hope that it will be useful, | 
|---|
| 18 | // but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|---|
| 19 | // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
|---|
| 20 | // GNU Library General Public License for more details. | 
|---|
| 21 | // | 
|---|
| 22 | // You should have received a copy of the GNU Library General Public License | 
|---|
| 23 | // along with the SC Toolkit; see the file COPYING.LIB.  If not, write to | 
|---|
| 24 | // the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. | 
|---|
| 25 | // | 
|---|
| 26 | // The U.S. Government is granted a limited license as per AL 91-7. | 
|---|
| 27 | // | 
|---|
| 28 |  | 
|---|
| 29 | #ifndef _util_group_memmtmpi_cc | 
|---|
| 30 | #define _util_group_memmtmpi_cc | 
|---|
| 31 |  | 
|---|
| 32 | #ifdef __GNUC__ | 
|---|
| 33 | #pragma implementation | 
|---|
| 34 | #endif | 
|---|
| 35 |  | 
|---|
| 36 | #include <assert.h> | 
|---|
| 37 |  | 
|---|
| 38 | #include <util/misc/formio.h> | 
|---|
| 39 | #include <util/group/messmpi.h> | 
|---|
| 40 | #include <util/group/memmtmpi.h> | 
|---|
| 41 |  | 
|---|
| 42 | #define MPICH_SKIP_MPICXX | 
|---|
| 43 | #include <mpi.h> | 
|---|
| 44 |  | 
|---|
| 45 | using namespace std; | 
|---|
| 46 |  | 
|---|
| 47 | // Define this to use immediate mode.  This was added added to work | 
|---|
| 48 | // around bugs in non-immediate mode optimizations in an MPI impl. | 
|---|
| 49 | #undef USE_IMMEDIATE_MODE | 
|---|
| 50 |  | 
|---|
| 51 | namespace sc { | 
|---|
| 52 |  | 
|---|
| 53 | static const int dbufsize = 32768; | 
|---|
| 54 |  | 
|---|
| 55 | /////////////////////////////////////////////////////////////////////// | 
|---|
| 56 | // The MTMPIThread class | 
|---|
| 57 |  | 
|---|
| 58 | class MTMPIThread: public Thread { | 
|---|
| 59 | private: | 
|---|
| 60 | MTMPIMemoryGrp *mem_; | 
|---|
| 61 | int req_tag_; | 
|---|
| 62 | int tag_; | 
|---|
| 63 | unsigned int nreq_recd_; | 
|---|
| 64 | double chunk[dbufsize]; | 
|---|
| 65 | public: | 
|---|
| 66 | MTMPIThread(MTMPIMemoryGrp *, int reqtype, int tag); | 
|---|
| 67 | void run(); | 
|---|
| 68 | int run_one(); | 
|---|
| 69 | unsigned int nreq_recd() { return nreq_recd_; } | 
|---|
| 70 | void set_nreq_recd(unsigned int val) { nreq_recd_ = val; } | 
|---|
| 71 | }; | 
|---|
| 72 |  | 
|---|
| 73 | MTMPIThread::MTMPIThread(MTMPIMemoryGrp *mem, | 
|---|
| 74 | int reqtype, int tag) | 
|---|
| 75 | { | 
|---|
| 76 | mem_ = mem; | 
|---|
| 77 | req_tag_ = reqtype; | 
|---|
| 78 | tag_ = tag; | 
|---|
| 79 | nreq_recd_ = 0; | 
|---|
| 80 | } | 
|---|
| 81 |  | 
|---|
| 82 | void | 
|---|
| 83 | MTMPIThread::run() | 
|---|
| 84 | { | 
|---|
| 85 | while(run_one()); | 
|---|
| 86 | } | 
|---|
| 87 |  | 
|---|
| 88 | int | 
|---|
| 89 | MTMPIThread::run_one() | 
|---|
| 90 | { | 
|---|
| 91 | int i; | 
|---|
| 92 | int dsize; | 
|---|
| 93 | int dremain; | 
|---|
| 94 | int doffset; | 
|---|
| 95 | long l; | 
|---|
| 96 | MemoryDataRequest req; | 
|---|
| 97 | MPI_Status status; | 
|---|
| 98 | #ifndef USE_IMMEDIATE_MODE | 
|---|
| 99 | MPI_Recv(req.data(),req.nbytes(),MPI_BYTE,MPI_ANY_SOURCE, | 
|---|
| 100 | req_tag_,mem_->comm_comm_,&status); | 
|---|
| 101 | #else | 
|---|
| 102 | MPI_Request mpireq; | 
|---|
| 103 | MPI_Irecv(req.data(),req.nbytes(),MPI_BYTE,MPI_ANY_SOURCE, | 
|---|
| 104 | req_tag_,mem_->comm_comm_,&mpireq); | 
|---|
| 105 | MPI_Wait(&mpireq,&status); | 
|---|
| 106 | #endif // USE_IMMEDIATE_MODE | 
|---|
| 107 | int rtag = req.serial_number(); | 
|---|
| 108 | if (mem_->debug_) { | 
|---|
| 109 | mem_->print_lock_->lock(); | 
|---|
| 110 | req.print("RECV",mem_->hout); | 
|---|
| 111 | mem_->print_lock_->unlock(); | 
|---|
| 112 | } | 
|---|
| 113 | if (req.touches_data()) { | 
|---|
| 114 | assert(req.size() >= 0); | 
|---|
| 115 | if (req.offset()+req.size() > mem_->localsize()) { | 
|---|
| 116 | mem_->print_lock_->lock(); | 
|---|
| 117 | req.print("BAD RECV"); | 
|---|
| 118 | ExEnv::outn() << "mem_->localsize() = " << mem_->localsize() << endl; | 
|---|
| 119 | mem_->print_lock_->lock(); | 
|---|
| 120 | } | 
|---|
| 121 | assert(req.offset()+req.size() <= mem_->localsize()); | 
|---|
| 122 | } | 
|---|
| 123 | switch (req.request()) { | 
|---|
| 124 | case MemoryDataRequest::Deactivate: | 
|---|
| 125 | return 0; | 
|---|
| 126 | case MemoryDataRequest::Retrieve: | 
|---|
| 127 | nreq_recd_++; | 
|---|
| 128 | if (req.lock()) | 
|---|
| 129 | mem_->obtain_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
| 130 | MPI_Send(&mem_->data_[req.offset()],req.size(),MPI_BYTE, | 
|---|
| 131 | req.node(),rtag,mem_->comp_comm_); | 
|---|
| 132 | break; | 
|---|
| 133 | case MemoryDataRequest::Replace: | 
|---|
| 134 | nreq_recd_++; | 
|---|
| 135 | // May be able to get rid of this MPI_Send - MLL | 
|---|
| 136 | MPI_Send(&tag_,1,MPI_INT,req.node(),rtag,mem_->comp_comm_); | 
|---|
| 137 | MPI_Recv(&mem_->data_[req.offset()],req.size(),MPI_BYTE, | 
|---|
| 138 | req.node(),tag_,mem_->comm_comm_,&status); | 
|---|
| 139 | if (req.lock()) | 
|---|
| 140 | mem_->release_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
| 141 | break; | 
|---|
| 142 | case MemoryDataRequest::DoubleSum: | 
|---|
| 143 | nreq_recd_++; | 
|---|
| 144 | //      MPI_Send(&tag_,1,MPI_INT,req.node(),rtag,mem_->comm_); | 
|---|
| 145 | dsize = req.size()/sizeof(double); | 
|---|
| 146 | dremain = dsize; | 
|---|
| 147 | doffset = req.offset()/sizeof(double); | 
|---|
| 148 | mem_->obtain_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
| 149 | while(dremain>0) { | 
|---|
| 150 | int dchunksize = dbufsize; | 
|---|
| 151 | if (dremain < dchunksize) dchunksize = dremain; | 
|---|
| 152 | #ifndef USE_IMMEDIATE_MODE | 
|---|
| 153 | MPI_Recv(chunk,dchunksize,MPI_DOUBLE, | 
|---|
| 154 | req.node(),rtag ,mem_->comm_comm_,&status); | 
|---|
| 155 | #else | 
|---|
| 156 | MPI_Request mpireq; | 
|---|
| 157 | MPI_Irecv(chunk,dchunksize,MPI_DOUBLE, | 
|---|
| 158 | req.node(),rtag ,mem_->comm_comm_,&mpireq); | 
|---|
| 159 | MPI_Wait(&mpireq,&status); | 
|---|
| 160 | #endif // USE_IMMEDIATE_MODE | 
|---|
| 161 | double *source_data = &((double*)mem_->data_)[doffset]; | 
|---|
| 162 | for (i=0; i<dchunksize; i++) { | 
|---|
| 163 | source_data[i] += chunk[i]; | 
|---|
| 164 | } | 
|---|
| 165 | dremain -= dchunksize; | 
|---|
| 166 | doffset += dchunksize; | 
|---|
| 167 | } | 
|---|
| 168 | mem_->release_local_lock(req.offset(), req.offset()+req.size()); | 
|---|
| 169 | break; | 
|---|
| 170 | default: | 
|---|
| 171 | mem_->print_lock_->lock(); | 
|---|
| 172 | ExEnv::outn() << "MTMPIThread: bad memory data request" << endl; | 
|---|
| 173 | mem_->print_lock_->unlock(); | 
|---|
| 174 | abort(); | 
|---|
| 175 | } | 
|---|
| 176 | return 1; | 
|---|
| 177 | } | 
|---|
| 178 |  | 
|---|
| 179 | /////////////////////////////////////////////////////////////////////// | 
|---|
| 180 | // The MTMPIMemoryGrp class | 
|---|
| 181 |  | 
|---|
| 182 | static ClassDesc MTMPIMemoryGrp_cd( | 
|---|
| 183 | typeid(MTMPIMemoryGrp),"MTMPIMemoryGrp",1,"public ActiveMsgMemoryGrp", | 
|---|
| 184 | 0, create<MTMPIMemoryGrp>, 0); | 
|---|
| 185 |  | 
|---|
| 186 | MTMPIMemoryGrp::MTMPIMemoryGrp(const Ref<MessageGrp>& msg, | 
|---|
| 187 | const Ref<ThreadGrp>& th, | 
|---|
| 188 | MPI_Comm comm): | 
|---|
| 189 | ActiveMsgMemoryGrp(msg) | 
|---|
| 190 | { | 
|---|
| 191 | if (debug_) ExEnv::outn() << "MTMPIMemoryGrp CTOR entered" << endl; | 
|---|
| 192 |  | 
|---|
| 193 | th_ = th; | 
|---|
| 194 |  | 
|---|
| 195 | init_mtmpimg(comm,th_->nthread()); | 
|---|
| 196 | } | 
|---|
| 197 |  | 
|---|
| 198 | MTMPIMemoryGrp::MTMPIMemoryGrp(const Ref<KeyVal>& keyval): | 
|---|
| 199 | ActiveMsgMemoryGrp(keyval) | 
|---|
| 200 | { | 
|---|
| 201 | if (debug_) ExEnv::outn() << "MTMPIMemoryGrp keyval CTOR entered" << endl; | 
|---|
| 202 |  | 
|---|
| 203 | th_ = ThreadGrp::get_default_threadgrp(); | 
|---|
| 204 |  | 
|---|
| 205 | KeyValValueint nthreaddef(th_->nthread()); | 
|---|
| 206 | int nthread = keyval->intvalue("num_threads",nthreaddef); | 
|---|
| 207 | ExEnv::out0() << indent << "MTMPIMemoryGrp: num_threads = " << nthread << endl; | 
|---|
| 208 |  | 
|---|
| 209 | init_mtmpimg(MPI_COMM_WORLD,nthread); | 
|---|
| 210 | } | 
|---|
| 211 |  | 
|---|
| 212 | MTMPIMemoryGrp::~MTMPIMemoryGrp() | 
|---|
| 213 | { | 
|---|
| 214 | deactivate(); | 
|---|
| 215 | for (int i=0; i<th_->nthread()-1; i++) { | 
|---|
| 216 | delete thread_[i]; | 
|---|
| 217 | } | 
|---|
| 218 | delete[] thread_; | 
|---|
| 219 | delete[] nreq_sent_; | 
|---|
| 220 | delete[] nreq_sent_buf_; | 
|---|
| 221 | } | 
|---|
| 222 |  | 
|---|
| 223 | void | 
|---|
| 224 | MTMPIMemoryGrp::init_mtmpimg(MPI_Comm comm, int nthread) | 
|---|
| 225 | { | 
|---|
| 226 | int i; | 
|---|
| 227 | active_ = 0; | 
|---|
| 228 |  | 
|---|
| 229 | if (nthread < 2) nthread = 2; | 
|---|
| 230 | th_ = th_->clone(nthread); | 
|---|
| 231 | nthread = th_->nthread(); | 
|---|
| 232 |  | 
|---|
| 233 | if (debug_) { | 
|---|
| 234 | char name[256]; | 
|---|
| 235 | sprintf(name, "mpqc.hand.%d", me()); | 
|---|
| 236 | hout.open(name); | 
|---|
| 237 | sprintf(name, "mpqc.main.%d", me()); | 
|---|
| 238 | mout.open(name); | 
|---|
| 239 | } | 
|---|
| 240 |  | 
|---|
| 241 | MPI_Comm_dup(comm, &comp_comm_); | 
|---|
| 242 | MPI_Comm_dup(comm, &comm_comm_); | 
|---|
| 243 |  | 
|---|
| 244 | MPI_Errhandler_set(comp_comm_, MPI_ERRORS_ARE_FATAL); | 
|---|
| 245 | MPI_Errhandler_set(comm_comm_, MPI_ERRORS_ARE_FATAL); | 
|---|
| 246 |  | 
|---|
| 247 | serial_ = 0; | 
|---|
| 248 | req_tag_ = 15001; | 
|---|
| 249 |  | 
|---|
| 250 | serial_lock_ = th_->new_lock(); | 
|---|
| 251 |  | 
|---|
| 252 | thread_ = new MTMPIThread*[nthread-1]; | 
|---|
| 253 | th_->add_thread(0,0); | 
|---|
| 254 | for (i=1; i<nthread; i++) { | 
|---|
| 255 | thread_[i-1] = new MTMPIThread(this,req_tag_,req_tag_ + i); | 
|---|
| 256 | th_->add_thread(i,thread_[i-1]); | 
|---|
| 257 | } | 
|---|
| 258 | print_lock_ = th_->new_lock(); | 
|---|
| 259 |  | 
|---|
| 260 | nreq_sent_ = new unsigned int[n()]; | 
|---|
| 261 | memset(nreq_sent_, 0, sizeof(unsigned int)*n()); | 
|---|
| 262 | nreq_sent_buf_ = new unsigned int[n()]; | 
|---|
| 263 | } | 
|---|
| 264 |  | 
|---|
| 265 | int | 
|---|
| 266 | MTMPIMemoryGrp::serial(int node) | 
|---|
| 267 | { | 
|---|
| 268 | serial_lock_->lock(); | 
|---|
| 269 | nreq_sent_[node]++; | 
|---|
| 270 | int r = serial_; | 
|---|
| 271 | serial_++; | 
|---|
| 272 | if (serial_ == req_tag_) serial_ = 0; | 
|---|
| 273 | serial_lock_->unlock(); | 
|---|
| 274 | return r; | 
|---|
| 275 | } | 
|---|
| 276 |  | 
|---|
| 277 | void | 
|---|
| 278 | MTMPIMemoryGrp::retrieve_data(void *data, int node, int offset, int size, | 
|---|
| 279 | int lock) | 
|---|
| 280 | { | 
|---|
| 281 | MemoryDataRequest req(MemoryDataRequest::Retrieve,me(),offset,size,lock, | 
|---|
| 282 | serial(node)); | 
|---|
| 283 | int tag = req.serial_number(); | 
|---|
| 284 |  | 
|---|
| 285 | // send the request | 
|---|
| 286 | if (debug_) { | 
|---|
| 287 | print_lock_->lock(); | 
|---|
| 288 | req.print("SEND",mout); | 
|---|
| 289 | print_lock_->unlock(); | 
|---|
| 290 | } | 
|---|
| 291 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_); | 
|---|
| 292 |  | 
|---|
| 293 | // receive the data | 
|---|
| 294 | MPI_Status status; | 
|---|
| 295 | MPI_Recv(data,size,MPI_BYTE,node,tag,comp_comm_,&status); | 
|---|
| 296 | } | 
|---|
| 297 |  | 
|---|
| 298 | void | 
|---|
| 299 | MTMPIMemoryGrp::replace_data(void *data, int node, int offset, int size, | 
|---|
| 300 | int unlock) | 
|---|
| 301 | { | 
|---|
| 302 | MemoryDataRequest req(MemoryDataRequest::Replace,me(),offset,size,unlock, | 
|---|
| 303 | serial(node)); | 
|---|
| 304 | int tag = req.serial_number(); | 
|---|
| 305 |  | 
|---|
| 306 | if (debug_) { | 
|---|
| 307 | print_lock_->lock(); | 
|---|
| 308 | req.print("SEND",mout); | 
|---|
| 309 | print_lock_->unlock(); | 
|---|
| 310 | } | 
|---|
| 311 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_); | 
|---|
| 312 |  | 
|---|
| 313 | // wait for the go ahead message | 
|---|
| 314 | MPI_Status status; | 
|---|
| 315 | int rtag; | 
|---|
| 316 | MPI_Recv(&rtag,1,MPI_INT,node,tag,comp_comm_,&status); | 
|---|
| 317 |  | 
|---|
| 318 | // send the data | 
|---|
| 319 | MPI_Send(data,size,MPI_BYTE,node,rtag,comm_comm_); | 
|---|
| 320 | } | 
|---|
| 321 |  | 
|---|
| 322 | void | 
|---|
| 323 | MTMPIMemoryGrp::sum_data(double *data, int node, int offset, int size) | 
|---|
| 324 | { | 
|---|
| 325 | MemoryDataRequest req(MemoryDataRequest::DoubleSum,me(),offset,size, | 
|---|
| 326 | 0, serial(node)); | 
|---|
| 327 | int tag = req.serial_number(); | 
|---|
| 328 |  | 
|---|
| 329 | // send the request | 
|---|
| 330 | if (debug_) { | 
|---|
| 331 | print_lock_->lock(); | 
|---|
| 332 | req.print("SEND",mout); | 
|---|
| 333 | print_lock_->unlock(); | 
|---|
| 334 | } | 
|---|
| 335 | #ifndef USE_IMMEDIATE_MODE | 
|---|
| 336 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_); | 
|---|
| 337 | #else | 
|---|
| 338 | MPI_Status status; | 
|---|
| 339 | MPI_Request mpireq; | 
|---|
| 340 | MPI_Isend(req.data(),req.nbytes(),MPI_BYTE,node,req_tag_,comm_comm_,&mpireq); | 
|---|
| 341 | MPI_Wait(&mpireq,&status); | 
|---|
| 342 | #endif // USE_IMMEDIATE_MODE | 
|---|
| 343 |  | 
|---|
| 344 | // wait for the go ahead message | 
|---|
| 345 | //  MPI_Status status; | 
|---|
| 346 | //  int rtag; | 
|---|
| 347 | //  MPI_Recv(&rtag,1,MPI_INT,node,tag,comm_,&status); | 
|---|
| 348 |  | 
|---|
| 349 | int dsize = size/sizeof(double); | 
|---|
| 350 | int dremain = dsize; | 
|---|
| 351 | int dcurrent = 0; | 
|---|
| 352 | while(dremain>0) { | 
|---|
| 353 | int dchunksize = dbufsize; | 
|---|
| 354 | if (dremain < dchunksize) dchunksize = dremain; | 
|---|
| 355 | // send the data | 
|---|
| 356 | #ifndef USE_IMMEDIATE_MODE | 
|---|
| 357 | MPI_Send(&data[dcurrent],dchunksize,MPI_DOUBLE, | 
|---|
| 358 | node,tag,comm_comm_); | 
|---|
| 359 | #else | 
|---|
| 360 | MPI_Request mpireq; | 
|---|
| 361 | MPI_Isend(&data[dcurrent],dchunksize,MPI_DOUBLE, | 
|---|
| 362 | node,tag,comm_comm_,&mpireq); | 
|---|
| 363 | MPI_Wait(&mpireq,&status); | 
|---|
| 364 | #endif // USE_IMMEDIATE_MODE | 
|---|
| 365 | dcurrent += dchunksize; | 
|---|
| 366 | dremain -= dchunksize; | 
|---|
| 367 | } | 
|---|
| 368 | } | 
|---|
| 369 |  | 
|---|
| 370 | void | 
|---|
| 371 | MTMPIMemoryGrp::activate() | 
|---|
| 372 | { | 
|---|
| 373 | // Only remote requests require the handler.  There are only remote | 
|---|
| 374 | // requests if there is more than one node. | 
|---|
| 375 | if (n() == 1) return; | 
|---|
| 376 |  | 
|---|
| 377 | if (th_->nthread() < 2) { | 
|---|
| 378 | ExEnv::outn() << "MTMPIMemoryGrp didn't get enough threads" << endl; | 
|---|
| 379 | abort(); | 
|---|
| 380 | } | 
|---|
| 381 |  | 
|---|
| 382 | if (active_) return; | 
|---|
| 383 | active_ = 1; | 
|---|
| 384 |  | 
|---|
| 385 | th_->start_threads(); | 
|---|
| 386 | } | 
|---|
| 387 |  | 
|---|
| 388 | void | 
|---|
| 389 | MTMPIMemoryGrp::deactivate() | 
|---|
| 390 | { | 
|---|
| 391 | if (!active_) return; | 
|---|
| 392 | active_ = 0; | 
|---|
| 393 |  | 
|---|
| 394 | // send a shutdown message | 
|---|
| 395 | MemoryDataRequest req(MemoryDataRequest::Deactivate); | 
|---|
| 396 | if (debug_) { | 
|---|
| 397 | print_lock_->lock(); | 
|---|
| 398 | req.print("SEND",mout); | 
|---|
| 399 | print_lock_->unlock(); | 
|---|
| 400 | } | 
|---|
| 401 | for (int i=1; i<th_->nthread(); i++) { | 
|---|
| 402 | #ifndef USE_IMMEDIATE_MODE | 
|---|
| 403 | MPI_Send(req.data(),req.nbytes(),MPI_BYTE,me(),req_tag_,comm_comm_); | 
|---|
| 404 | #else | 
|---|
| 405 | MPI_Request mpireq; | 
|---|
| 406 | MPI_Status status; | 
|---|
| 407 | MPI_Isend(req.data(),req.nbytes(),MPI_BYTE,me(),req_tag_,comm_comm_,&mpireq); | 
|---|
| 408 | MPI_Wait(&mpireq,&status); | 
|---|
| 409 | #endif // USE_IMMEDIATE_MODE | 
|---|
| 410 | } | 
|---|
| 411 |  | 
|---|
| 412 | // wait on the thread to shutdown | 
|---|
| 413 | th_->wait_threads(); | 
|---|
| 414 | } | 
|---|
| 415 |  | 
|---|
| 416 | void | 
|---|
| 417 | MTMPIMemoryGrp::sync() | 
|---|
| 418 | { | 
|---|
| 419 | if (active_) { | 
|---|
| 420 | MPI_Allreduce(nreq_sent_, nreq_sent_buf_, | 
|---|
| 421 | n(), MPI_UNSIGNED, MPI_SUM, comm_comm_); | 
|---|
| 422 | deactivate(); | 
|---|
| 423 | unsigned int nreq_recd = 0; | 
|---|
| 424 | for (int i=0; i<th_->nthread()-1; i++) { | 
|---|
| 425 | nreq_recd += thread_[i]->nreq_recd(); | 
|---|
| 426 | thread_[i]->set_nreq_recd(0); | 
|---|
| 427 | } | 
|---|
| 428 | int n_outstanding = nreq_sent_buf_[me()] - nreq_recd; | 
|---|
| 429 | for (int i=0; i<n_outstanding; i++) { | 
|---|
| 430 | thread_[0]->run_one(); | 
|---|
| 431 | } | 
|---|
| 432 | memset(nreq_sent_, 0, sizeof(unsigned int)*n()); | 
|---|
| 433 | // Make sure processing of all outstanding requests is finished | 
|---|
| 434 | // before starting the next phase. | 
|---|
| 435 | MPI_Barrier(comm_comm_); | 
|---|
| 436 | activate(); | 
|---|
| 437 | } | 
|---|
| 438 | else { | 
|---|
| 439 | MPI_Barrier(comm_comm_); | 
|---|
| 440 | } | 
|---|
| 441 | } | 
|---|
| 442 |  | 
|---|
| 443 | #endif | 
|---|
| 444 |  | 
|---|
| 445 | ///////////////////////////////////////////////////////////////////////////// | 
|---|
| 446 |  | 
|---|
| 447 | } | 
|---|
| 448 |  | 
|---|
| 449 | // Local Variables: | 
|---|
| 450 | // mode: c++ | 
|---|
| 451 | // c-file-style: "CLJ" | 
|---|
| 452 | // End: | 
|---|