|
| MPI_int | Bcast (void *buf, MPI_int num, MPI_Datatype type, MPI_int source_rank, MPI_Comm comm) |
| | dumb wrapper
|
| |
| MPI_int | Alltoall (void *send, MPI_int sendNum, MPI_Datatype typeSend, void *recv, MPI_int recvNum, MPI_Datatype typeRecv, MPI_Comm comm) |
| | Wrapper over MPI_Alltoall (fixed per-peer count).
|
| |
| MPI_int | Alltoallv (void *send, MPI_int *sendSizes, MPI_int *sendStarts, MPI_Datatype sendType, void *recv, MPI_int *recvSizes, MPI_int *recvStarts, MPI_Datatype recvType, MPI_Comm comm) |
| | Wrapper over MPI_Alltoallv (variable per-peer counts + displacements).
|
| |
| MPI_int | Allreduce (const void *sendbuf, void *recvbuf, MPI_int count, MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) |
| | Wrapper over MPI_Allreduce.
|
| |
| MPI_int | Scan (const void *sendbuf, void *recvbuf, MPI_int count, MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) |
| | Wrapper over MPI_Scan (inclusive prefix reduction).
|
| |
| MPI_int | Allgather (const void *sendbuf, MPI_int sendcount, MPI_Datatype sendtype, void *recvbuf, MPI_int recvcount, MPI_Datatype recvtype, MPI_Comm comm) |
| | Wrapper over MPI_Allgather.
|
| |
| MPI_int | Barrier (MPI_Comm comm) |
| | Wrapper over MPI_Barrier.
|
| |
| MPI_int | BarrierLazy (MPI_Comm comm, uint64_t checkNanoSecs) |
| | Polling barrier that sleeps checkNanoSecs ns between MPI_Test calls. Reduces CPU spin when many ranks wait unevenly.
|
| |
| MPI_int | WaitallLazy (MPI_int count, MPI_Request *reqs, MPI_Status *statuses, uint64_t checkNanoSecs=10000000) |
| | Like WaitallAuto but sleeps checkNanoSecs ns between polls.
|
| |
| MPI_int | WaitallAuto (MPI_int count, MPI_Request *reqs, MPI_Status *statuses) |
| | Wait on an array of requests, choosing between MPI_Waitall and the lazy-poll variant based on CommStrategy settings.
|
| |
| bool | isCudaAware () |
| | Runtime probe: is the current MPI implementation configured with CUDA-aware support? Affects whether arrays are transferred on-device or via the host round-trip.
|
| |
| int | GetMPIThreadLevel () |
| | Return the MPI thread-support level the current process was initialised with.
|
| |
| MPI_int | Init_thread (int *argc, char ***argv) |
| | Initialise MPI with thread support, honouring the DNDS_DISABLE_ASYNC_MPI environment override.
|
| |
| int | Finalize () |
| | Release DNDSR-registered MPI resources then call MPI_Finalize.
|
| |
| void | AllreduceOneReal (real &v, MPI_Op op, const MPIInfo &mpi) |
| | Single-scalar Allreduce helper for reals (in-place, count = 1).
|
| |
| void | AllreduceOneIndex (index &v, MPI_Op op, const MPIInfo &mpi) |
| | Single-scalar Allreduce helper for indices (in-place, count = 1).
|
| |
| void | pybind11_Init_thread (py::module_ &m) |
| |
| void | pybind11_MPI_Operations (py::module_ &m) |
| |