File: simpleMPI.cpp

package info (click to toggle)
nvidia-cuda-samples 12.4.1~dfsg-1
  • links: PTS, VCS
  • area: contrib
  • in suites: forky, sid, trixie
  • size: 313,216 kB
  • sloc: cpp: 82,042; makefile: 53,971; xml: 15,381; ansic: 8,630; sh: 91; python: 74
file content (128 lines) | stat: -rw-r--r-- 4,230 bytes parent folder | download | duplicates (2)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
/* Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 *  * Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 *  * Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 *  * Neither the name of NVIDIA CORPORATION nor the names of its
 *    contributors may be used to endorse or promote products derived
 *    from this software without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY
 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
 * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE COPYRIGHT OWNER OR
 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
 * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 */

/* Simple example demonstrating how to use MPI with CUDA
*
*  Generate some random numbers on one node.
*  Dispatch them to all nodes.
*  Compute their square root on each node's GPU.
*  Compute the average of the results using MPI.
*
*  simpleMPI.cpp: main program, compiled with mpicxx on linux/Mac platforms
*                 on Windows, please download the Microsoft HPC Pack SDK 2008
*/

// MPI include
#include <mpi.h>

// System includes
#include <iostream>

using std::cout;
using std::cerr;
using std::endl;

// User include
#include "simpleMPI.h"

// Error handling macros
#define MPI_CHECK(call)                          \
  if ((call) != MPI_SUCCESS) {                   \
    cerr << "MPI error calling \"" #call "\"\n"; \
    my_abort(-1);                                \
  }

// Host code
// No CUDA here, only MPI
int main(int argc, char *argv[]) {
  // Dimensions of the dataset
  int blockSize = 256;
  int gridSize = 10000;
  int dataSizePerNode = gridSize * blockSize;

  // Initialize MPI state
  MPI_CHECK(MPI_Init(&argc, &argv));

  // Get our MPI node number and node count
  int commSize, commRank;
  MPI_CHECK(MPI_Comm_size(MPI_COMM_WORLD, &commSize));
  MPI_CHECK(MPI_Comm_rank(MPI_COMM_WORLD, &commRank));

  // Generate some random numbers on the root node (node 0)
  int dataSizeTotal = dataSizePerNode * commSize;
  float *dataRoot = NULL;

  // Are we the root node?
  if (commRank == 0) {
    cout << "Running on " << commSize << " nodes" << endl;
    dataRoot = new float[dataSizeTotal];
    initData(dataRoot, dataSizeTotal);
  }

  // Allocate a buffer on each node
  float *dataNode = new float[dataSizePerNode];

  // Dispatch a portion of the input data to each node
  MPI_CHECK(MPI_Scatter(dataRoot, dataSizePerNode, MPI_FLOAT, dataNode,
                        dataSizePerNode, MPI_FLOAT, 0, MPI_COMM_WORLD));

  if (commRank == 0) {
    // No need for root data any more
    delete[] dataRoot;
  }

  // On each node, run computation on GPU
  computeGPU(dataNode, blockSize, gridSize);

  // Reduction to the root node, computing the sum of output elements
  float sumNode = sum(dataNode, dataSizePerNode);
  float sumRoot;

  MPI_CHECK(
      MPI_Reduce(&sumNode, &sumRoot, 1, MPI_FLOAT, MPI_SUM, 0, MPI_COMM_WORLD));

  if (commRank == 0) {
    float average = sumRoot / dataSizeTotal;
    cout << "Average of square roots is: " << average << endl;
  }

  // Cleanup
  delete[] dataNode;
  MPI_CHECK(MPI_Finalize());

  if (commRank == 0) {
    cout << "PASSED\n";
  }

  return 0;
}

// Shut down MPI cleanly if something goes wrong
void my_abort(int err) {
  cout << "Test FAILED\n";
  MPI_Abort(MPI_COMM_WORLD, err);
}