mirror of
https://github.com/NGSolve/netgen.git
synced 2024-11-11 16:49:16 +05:00
MPI unification
This commit is contained in:
parent
b62c7b30ef
commit
fd77d17e2b
@ -7,8 +7,10 @@
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
#include "array.hpp"
|
#include "array.hpp"
|
||||||
|
#include "table.hpp"
|
||||||
#include "exception.hpp"
|
#include "exception.hpp"
|
||||||
#include "profiler.hpp"
|
#include "profiler.hpp"
|
||||||
|
#include "ngstream.hpp"
|
||||||
|
|
||||||
namespace ngcore
|
namespace ngcore
|
||||||
{
|
{
|
||||||
@ -52,6 +54,23 @@ namespace ngcore
|
|||||||
return GetMPIType<T>();
|
return GetMPIType<T>();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
inline void MyMPI_WaitAll (FlatArray<MPI_Request> requests)
|
||||||
|
{
|
||||||
|
static Timer t("MPI - WaitAll"); RegionTimer reg(t);
|
||||||
|
if (!requests.Size()) return;
|
||||||
|
MPI_Waitall (requests.Size(), requests.Data(), MPI_STATUSES_IGNORE);
|
||||||
|
}
|
||||||
|
|
||||||
|
inline int MyMPI_WaitAny (FlatArray<MPI_Request> requests)
|
||||||
|
{
|
||||||
|
int nr;
|
||||||
|
MPI_Waitany (requests.Size(), requests.Data(), &nr, MPI_STATUS_IGNORE);
|
||||||
|
return nr;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class NgMPI_Comm
|
class NgMPI_Comm
|
||||||
{
|
{
|
||||||
protected:
|
protected:
|
||||||
@ -265,11 +284,25 @@ namespace ngcore
|
|||||||
|
|
||||||
template <typename T, typename T2 = decltype(GetMPIType<T>())>
|
template <typename T, typename T2 = decltype(GetMPIType<T>())>
|
||||||
void Bcast (T & s, int root = 0) const {
|
void Bcast (T & s, int root = 0) const {
|
||||||
if (size == 1) return ;
|
if (size == 1) return;
|
||||||
static Timer t("MPI - Bcast"); RegionTimer reg(t);
|
static Timer t("MPI - Bcast"); RegionTimer reg(t);
|
||||||
MPI_Bcast (&s, 1, GetMPIType<T>(), root, comm);
|
MPI_Bcast (&s, 1, GetMPIType<T>(), root, comm);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
template <class T>
|
||||||
|
void Bcast (Array<T> & d, int root = 0)
|
||||||
|
{
|
||||||
|
if (size == 1) return;
|
||||||
|
|
||||||
|
int ds = d.Size();
|
||||||
|
Bcast (ds, root);
|
||||||
|
if (Rank() != root) d.SetSize (ds);
|
||||||
|
if (ds != 0)
|
||||||
|
MPI_Bcast (d.Data(), ds, GetMPIType<T>(), root, comm);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
void Bcast (std::string & s, int root = 0) const
|
void Bcast (std::string & s, int root = 0) const
|
||||||
{
|
{
|
||||||
if (size == 1) return;
|
if (size == 1) return;
|
||||||
@ -335,6 +368,37 @@ namespace ngcore
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
template <typename T>
|
||||||
|
void ExchangeTable (DynamicTable<T> & send_data,
|
||||||
|
DynamicTable<T> & recv_data, int tag)
|
||||||
|
{
|
||||||
|
Array<int> send_sizes(size);
|
||||||
|
Array<int> recv_sizes(size);
|
||||||
|
|
||||||
|
for (int i = 0; i < size; i++)
|
||||||
|
send_sizes[i] = send_data[i].Size();
|
||||||
|
|
||||||
|
AllToAll (send_sizes, recv_sizes);
|
||||||
|
|
||||||
|
recv_data = DynamicTable<T> (recv_sizes, true);
|
||||||
|
|
||||||
|
Array<MPI_Request> requests;
|
||||||
|
for (int dest = 0; dest < size; dest++)
|
||||||
|
if (dest != rank && send_data[dest].Size())
|
||||||
|
requests.Append (ISend (FlatArray<T>(send_data[dest]), dest, tag));
|
||||||
|
|
||||||
|
for (int dest = 0; dest < size; dest++)
|
||||||
|
if (dest != rank && recv_data[dest].Size())
|
||||||
|
requests.Append (IRecv (FlatArray<T>(recv_data[dest]), dest, tag));
|
||||||
|
|
||||||
|
MyMPI_WaitAll (requests);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
NgMPI_Comm SubCommunicator (FlatArray<int> procs) const
|
NgMPI_Comm SubCommunicator (FlatArray<int> procs) const
|
||||||
{
|
{
|
||||||
MPI_Comm subcomm;
|
MPI_Comm subcomm;
|
||||||
@ -347,19 +411,44 @@ namespace ngcore
|
|||||||
|
|
||||||
}; // class NgMPI_Comm
|
}; // class NgMPI_Comm
|
||||||
|
|
||||||
inline void MyMPI_WaitAll (FlatArray<MPI_Request> requests)
|
|
||||||
{
|
|
||||||
static Timer t("MPI - WaitAll"); RegionTimer reg(t);
|
|
||||||
if (!requests.Size()) return;
|
|
||||||
MPI_Waitall (requests.Size(), requests.Data(), MPI_STATUSES_IGNORE);
|
|
||||||
}
|
|
||||||
|
|
||||||
inline int MyMPI_WaitAny (FlatArray<MPI_Request> requests)
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
class MyMPI
|
||||||
{
|
{
|
||||||
int nr;
|
bool initialized_by_me;
|
||||||
MPI_Waitany (requests.Size(), requests.Data(), &nr, MPI_STATUS_IGNORE);
|
public:
|
||||||
return nr;
|
MyMPI(int argc, char ** argv)
|
||||||
}
|
{
|
||||||
|
int is_init = -1;
|
||||||
|
MPI_Initialized(&is_init);
|
||||||
|
if (!is_init)
|
||||||
|
{
|
||||||
|
MPI_Init (&argc, &argv);
|
||||||
|
initialized_by_me = true;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
initialized_by_me = false;
|
||||||
|
|
||||||
|
NgMPI_Comm comm(MPI_COMM_WORLD);
|
||||||
|
NGSOStream::SetGlobalActive (comm.Rank() == 0);
|
||||||
|
|
||||||
|
if (comm.Size() > 1)
|
||||||
|
TaskManager::SetNumThreads (1);
|
||||||
|
}
|
||||||
|
|
||||||
|
~MyMPI()
|
||||||
|
{
|
||||||
|
if (initialized_by_me)
|
||||||
|
MPI_Finalize ();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
#else // PARALLEL
|
#else // PARALLEL
|
||||||
class MPI_Comm {
|
class MPI_Comm {
|
||||||
@ -429,6 +518,14 @@ namespace ngcore
|
|||||||
template <typename T>
|
template <typename T>
|
||||||
void Bcast (T & s, int root = 0) const { ; }
|
void Bcast (T & s, int root = 0) const { ; }
|
||||||
|
|
||||||
|
template <class T>
|
||||||
|
void Bcast (Array<T> & d, int root = 0) { ; }
|
||||||
|
|
||||||
|
template <typename T>
|
||||||
|
void ExchangeTable (DynamicTable<T> & send_data,
|
||||||
|
DynamicTable<T> & recv_data, int tag) { ; }
|
||||||
|
|
||||||
|
|
||||||
NgMPI_Comm SubCommunicator (FlatArray<int> procs) const
|
NgMPI_Comm SubCommunicator (FlatArray<int> procs) const
|
||||||
{ return *this; }
|
{ return *this; }
|
||||||
};
|
};
|
||||||
@ -436,6 +533,13 @@ namespace ngcore
|
|||||||
inline void MyMPI_WaitAll (FlatArray<MPI_Request> requests) { ; }
|
inline void MyMPI_WaitAll (FlatArray<MPI_Request> requests) { ; }
|
||||||
inline int MyMPI_WaitAny (FlatArray<MPI_Request> requests) { return 0; }
|
inline int MyMPI_WaitAny (FlatArray<MPI_Request> requests) { return 0; }
|
||||||
|
|
||||||
|
class MyMPI
|
||||||
|
{
|
||||||
|
public:
|
||||||
|
MyMPI(int argc, char ** argv) { ; }
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
#endif // PARALLEL
|
#endif // PARALLEL
|
||||||
|
|
||||||
} // namespace ngcore
|
} // namespace ngcore
|
||||||
|
@ -465,7 +465,7 @@ namespace ngcore
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Creates table with a priori fixed entry sizes.
|
/// Creates table with a priori fixed entry sizes.
|
||||||
DynamicTable (const Array<int, IndexType> & entrysizes)
|
DynamicTable (const Array<int, IndexType> & entrysizes, bool setentrysize=false)
|
||||||
: data(entrysizes.Size())
|
: data(entrysizes.Size())
|
||||||
{
|
{
|
||||||
size_t cnt = 0;
|
size_t cnt = 0;
|
||||||
@ -479,7 +479,10 @@ namespace ngcore
|
|||||||
for (auto i : data.Range())
|
for (auto i : data.Range())
|
||||||
{
|
{
|
||||||
data[i].maxsize = entrysizes[i];
|
data[i].maxsize = entrysizes[i];
|
||||||
data[i].size = 0;
|
if (setentrysize)
|
||||||
|
data[i].size = entrysizes[i];
|
||||||
|
else
|
||||||
|
data[i].size = 0;
|
||||||
data[i].col = &oneblock[cnt];
|
data[i].col = &oneblock[cnt];
|
||||||
cnt += entrysizes[i];
|
cnt += entrysizes[i];
|
||||||
}
|
}
|
||||||
|
@ -26,7 +26,10 @@
|
|||||||
namespace ngcore
|
namespace ngcore
|
||||||
{
|
{
|
||||||
// MPI rank, nranks TODO: Rename
|
// MPI rank, nranks TODO: Rename
|
||||||
extern NGCORE_API int id, ntasks;
|
// [[deprecated("don't use global id/ntasks")]]
|
||||||
|
extern NGCORE_API int id;
|
||||||
|
// [[deprecated("don't use global id/ntasks")]]
|
||||||
|
extern NGCORE_API int ntasks;
|
||||||
|
|
||||||
NGCORE_API std::string Demangle(const char* typeinfo);
|
NGCORE_API std::string Demangle(const char* typeinfo);
|
||||||
|
|
||||||
|
@ -186,7 +186,7 @@ namespace netgen
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
template <typename T>
|
template <typename T>
|
||||||
// [[deprecated("do we need that ? ")]]
|
[[deprecated("do we need that ? ")]]
|
||||||
inline void MyMPI_ExchangeTable (TABLE<T> & send_data,
|
inline void MyMPI_ExchangeTable (TABLE<T> & send_data,
|
||||||
TABLE<T> & recv_data, int tag,
|
TABLE<T> & recv_data, int tag,
|
||||||
const NgMPI_Comm & comm)
|
const NgMPI_Comm & comm)
|
||||||
@ -217,6 +217,41 @@ namespace netgen
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
template <typename T>
|
||||||
|
[[deprecated("do we need that ? ")]]
|
||||||
|
inline void MyMPI_ExchangeTable (DynamicTable<T> & send_data,
|
||||||
|
DynamicTable<T> & recv_data, int tag,
|
||||||
|
const NgMPI_Comm & comm)
|
||||||
|
{
|
||||||
|
int rank = comm.Rank();
|
||||||
|
int ntasks = comm.Size();
|
||||||
|
|
||||||
|
Array<int> send_sizes(ntasks);
|
||||||
|
Array<int> recv_sizes(ntasks);
|
||||||
|
for (int i = 0; i < ntasks; i++)
|
||||||
|
send_sizes[i] = send_data[i].Size();
|
||||||
|
|
||||||
|
comm.AllToAll (send_sizes, recv_sizes);
|
||||||
|
|
||||||
|
// for (int i = 0; i < ntasks; i++)
|
||||||
|
// recv_data.SetEntrySize (i, recv_sizes[i], sizeof(T));
|
||||||
|
recv_data = DynamicTable<T> (recv_sizes, true);
|
||||||
|
|
||||||
|
Array<MPI_Request> requests;
|
||||||
|
for (int dest = 0; dest < ntasks; dest++)
|
||||||
|
if (dest != rank && send_data[dest].Size())
|
||||||
|
requests.Append (comm.ISend (FlatArray<T>(send_data[dest]), dest, tag));
|
||||||
|
|
||||||
|
for (int dest = 0; dest < ntasks; dest++)
|
||||||
|
if (dest != rank && recv_data[dest].Size())
|
||||||
|
requests.Append (comm.IRecv (FlatArray<T>(recv_data[dest]), dest, tag));
|
||||||
|
|
||||||
|
MyMPI_WaitAll (requests);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
[[deprecated("do we still send commands?")]]
|
[[deprecated("do we still send commands?")]]
|
||||||
extern void MyMPI_SendCmd (const char * cmd);
|
extern void MyMPI_SendCmd (const char * cmd);
|
||||||
[[deprecated("do we still send commands?")]]
|
[[deprecated("do we still send commands?")]]
|
||||||
@ -224,14 +259,14 @@ namespace netgen
|
|||||||
|
|
||||||
|
|
||||||
template <class T>
|
template <class T>
|
||||||
// [[deprecated("use comm.BCast instead")]]
|
[[deprecated("use comm.BCast instead")]]
|
||||||
inline void MyMPI_Bcast (T & s, MPI_Comm comm)
|
inline void MyMPI_Bcast (T & s, MPI_Comm comm)
|
||||||
{
|
{
|
||||||
MPI_Bcast (&s, 1, GetMPIType<T>(), 0, comm);
|
MPI_Bcast (&s, 1, GetMPIType<T>(), 0, comm);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class T>
|
template <class T>
|
||||||
// [[deprecated("use comm.BCast instead")]]
|
[[deprecated("use comm.BCast instead")]]
|
||||||
inline void MyMPI_Bcast (NgArray<T, 0> & s, NgMPI_Comm comm)
|
inline void MyMPI_Bcast (NgArray<T, 0> & s, NgMPI_Comm comm)
|
||||||
{
|
{
|
||||||
int size = s.Size();
|
int size = s.Size();
|
||||||
|
@ -10,7 +10,6 @@ namespace netgen
|
|||||||
|
|
||||||
// bool rational = true;
|
// bool rational = true;
|
||||||
|
|
||||||
|
|
||||||
static void ComputeGaussRule (int n, NgArray<double> & xi, NgArray<double> & wi)
|
static void ComputeGaussRule (int n, NgArray<double> & xi, NgArray<double> & wi)
|
||||||
{
|
{
|
||||||
xi.SetSize (n);
|
xi.SetSize (n);
|
||||||
@ -640,8 +639,9 @@ namespace netgen
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
#ifdef PARALLEL
|
// #ifdef PARALLEL
|
||||||
TABLE<int> send_orders(ntasks), recv_orders(ntasks);
|
// TABLE<int> send_orders(ntasks), recv_orders(ntasks);
|
||||||
|
DynamicTable<int> send_orders(ntasks), recv_orders(ntasks);
|
||||||
|
|
||||||
if (ntasks > 1 && working)
|
if (ntasks > 1 && working)
|
||||||
{
|
{
|
||||||
@ -654,11 +654,12 @@ namespace netgen
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (ntasks > 1)
|
if (ntasks > 1)
|
||||||
MyMPI_ExchangeTable (send_orders, recv_orders, MPI_TAG_CURVE, comm);
|
// MyMPI_ExchangeTable (send_orders, recv_orders, MPI_TAG_CURVE, comm);
|
||||||
|
comm.ExchangeTable (send_orders, recv_orders, MPI_TAG_CURVE);
|
||||||
|
|
||||||
if (ntasks > 1 && working)
|
if (ntasks > 1 && working)
|
||||||
{
|
{
|
||||||
NgArray<int> cnt(ntasks);
|
Array<int> cnt(ntasks);
|
||||||
cnt = 0;
|
cnt = 0;
|
||||||
for (int e = 0; e < edgeorder.Size(); e++)
|
for (int e = 0; e < edgeorder.Size(); e++)
|
||||||
for (auto proc : partop.GetDistantEdgeProcs(e))
|
for (auto proc : partop.GetDistantEdgeProcs(e))
|
||||||
@ -667,7 +668,7 @@ namespace netgen
|
|||||||
for (auto proc : partop.GetDistantFaceProcs(f))
|
for (auto proc : partop.GetDistantFaceProcs(f))
|
||||||
faceorder[f] = max(faceorder[f], recv_orders[proc][cnt[proc]++]);
|
faceorder[f] = max(faceorder[f], recv_orders[proc][cnt[proc]++]);
|
||||||
}
|
}
|
||||||
#endif
|
// #endif
|
||||||
|
|
||||||
|
|
||||||
edgecoeffsindex.SetSize (nedges+1);
|
edgecoeffsindex.SetSize (nedges+1);
|
||||||
@ -750,7 +751,9 @@ namespace netgen
|
|||||||
if (ntasks > 1)
|
if (ntasks > 1)
|
||||||
{
|
{
|
||||||
// distribute it ...
|
// distribute it ...
|
||||||
TABLE<double> senddata(ntasks), recvdata(ntasks);
|
// TABLE<double> senddata(ntasks), recvdata(ntasks);
|
||||||
|
DynamicTable<double> senddata(ntasks), recvdata(ntasks);
|
||||||
|
|
||||||
if (working)
|
if (working)
|
||||||
for (int e = 0; e < nedges; e++)
|
for (int e = 0; e < nedges; e++)
|
||||||
for (int proc : partop.GetDistantEdgeProcs(e))
|
for (int proc : partop.GetDistantEdgeProcs(e))
|
||||||
@ -767,7 +770,8 @@ namespace netgen
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MyMPI_ExchangeTable (senddata, recvdata, MPI_TAG_CURVE, comm);
|
// MyMPI_ExchangeTable (senddata, recvdata, MPI_TAG_CURVE, comm);
|
||||||
|
comm.ExchangeTable (senddata, recvdata, MPI_TAG_CURVE);
|
||||||
|
|
||||||
NgArray<int> cnt(ntasks);
|
NgArray<int> cnt(ntasks);
|
||||||
cnt = 0;
|
cnt = 0;
|
||||||
@ -945,7 +949,9 @@ namespace netgen
|
|||||||
if (ntasks > 1)
|
if (ntasks > 1)
|
||||||
{
|
{
|
||||||
// distribute it ...
|
// distribute it ...
|
||||||
TABLE<double> senddata(ntasks), recvdata(ntasks);
|
// TABLE<double> senddata(ntasks), recvdata(ntasks);
|
||||||
|
DynamicTable<double> senddata(ntasks), recvdata(ntasks);
|
||||||
|
|
||||||
if (working)
|
if (working)
|
||||||
for (int e = 0; e < nedges; e++)
|
for (int e = 0; e < nedges; e++)
|
||||||
for (int proc : partop.GetDistantEdgeProcs(e))
|
for (int proc : partop.GetDistantEdgeProcs(e))
|
||||||
@ -969,7 +975,9 @@ namespace netgen
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MyMPI_ExchangeTable (senddata, recvdata, MPI_TAG_CURVE, comm);
|
// MyMPI_ExchangeTable (senddata, recvdata, MPI_TAG_CURVE, comm);
|
||||||
|
comm.ExchangeTable (senddata, recvdata, MPI_TAG_CURVE);
|
||||||
|
|
||||||
NgArray<int> cnt(ntasks);
|
NgArray<int> cnt(ntasks);
|
||||||
cnt = 0;
|
cnt = 0;
|
||||||
if (working)
|
if (working)
|
||||||
@ -1145,7 +1153,8 @@ namespace netgen
|
|||||||
mesh.GetFaceDescriptor(mesh[i].GetIndex()).SurfNr();
|
mesh.GetFaceDescriptor(mesh[i].GetIndex()).SurfNr();
|
||||||
|
|
||||||
#ifdef PARALLEL
|
#ifdef PARALLEL
|
||||||
TABLE<int> send_surfnr(ntasks), recv_surfnr(ntasks);
|
// TABLE<int> send_surfnr(ntasks), recv_surfnr(ntasks);
|
||||||
|
DynamicTable<int> send_surfnr(ntasks), recv_surfnr(ntasks);
|
||||||
|
|
||||||
if (ntasks > 1 && working)
|
if (ntasks > 1 && working)
|
||||||
{
|
{
|
||||||
@ -1155,7 +1164,8 @@ namespace netgen
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (ntasks > 1)
|
if (ntasks > 1)
|
||||||
MyMPI_ExchangeTable (send_surfnr, recv_surfnr, MPI_TAG_CURVE, comm);
|
// MyMPI_ExchangeTable (send_surfnr, recv_surfnr, MPI_TAG_CURVE, comm);
|
||||||
|
comm.ExchangeTable (send_surfnr, recv_surfnr, MPI_TAG_CURVE);
|
||||||
|
|
||||||
if (ntasks > 1 && working)
|
if (ntasks > 1 && working)
|
||||||
{
|
{
|
||||||
|
@ -701,7 +701,7 @@ DLL_HEADER void ExportNetgenMeshing(py::module &m)
|
|||||||
}
|
}
|
||||||
|
|
||||||
istream * infile = nullptr;
|
istream * infile = nullptr;
|
||||||
NgArray<char> buf; // for distributing geometry!
|
Array<char> buf; // for distributing geometry!
|
||||||
int strs;
|
int strs;
|
||||||
|
|
||||||
if( id == 0) {
|
if( id == 0) {
|
||||||
@ -796,13 +796,16 @@ DLL_HEADER void ExportNetgenMeshing(py::module &m)
|
|||||||
}
|
}
|
||||||
|
|
||||||
if(ntasks>1) {
|
if(ntasks>1) {
|
||||||
#ifdef PARALLEL
|
// #ifdef PARALLEL
|
||||||
/** Scatter the geometry-string (no dummy-implementation in mpi_interface) **/
|
/** Scatter the geometry-string (no dummy-implementation in mpi_interface) **/
|
||||||
|
/*
|
||||||
int strs = buf.Size();
|
int strs = buf.Size();
|
||||||
MyMPI_Bcast(strs, comm);
|
MyMPI_Bcast(strs, comm);
|
||||||
if(strs>0)
|
if(strs>0)
|
||||||
MyMPI_Bcast(buf, comm);
|
MyMPI_Bcast(buf, comm);
|
||||||
#endif
|
*/
|
||||||
|
comm.Bcast(buf);
|
||||||
|
// #endif
|
||||||
}
|
}
|
||||||
|
|
||||||
shared_ptr<NetgenGeometry> geo;
|
shared_ptr<NetgenGeometry> geo;
|
||||||
|
@ -13,7 +13,7 @@
|
|||||||
#ifdef PARALLEL
|
#ifdef PARALLEL
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
|
|
||||||
extern void ParallelRun();
|
// extern void ParallelRun();
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#include "../libsrc/interface/writeuser.hpp"
|
#include "../libsrc/interface/writeuser.hpp"
|
||||||
@ -281,7 +281,7 @@ int main(int argc, char ** argv)
|
|||||||
#ifdef PARALLEL
|
#ifdef PARALLEL
|
||||||
else
|
else
|
||||||
{
|
{
|
||||||
ParallelRun();
|
// ParallelRun();
|
||||||
MPI_Finalize();
|
MPI_Finalize();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1,3 +1,6 @@
|
|||||||
|
#ifdef OLDFILE
|
||||||
|
|
||||||
|
|
||||||
#ifdef PARALLEL
|
#ifdef PARALLEL
|
||||||
|
|
||||||
#include "dlfcn.h"
|
#include "dlfcn.h"
|
||||||
@ -353,4 +356,7 @@ void ParallelRun()
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
Loading…
Reference in New Issue
Block a user