1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182
|
!
! Copyright (C) 2013, Northwestern University and Argonne National Laboratory
! See COPYRIGHT notice in top-level directory.
!
! $Id$
!
! This example shows how to use PnetCDF flexible API, nf90mpi_put_var_all()
! to write a 2D 4-byte integer array in parallel. It first defines a netCDF
! variable of size global_nx * global_ny where
! global_nx == 5 and
! global_ny == (4 * number of MPI processes).
! The data partitioning pattern is a column-wise partitioning across all
! processes. Each process writes a subarray of size nx * ny.
! The local buffer has a ghost cell of length 3 surrounding the 2D array
! integer buf(nx+2*ghost_len, ny+2*ghost_len)
! Note the description above follows the Fortran array index order.
!
! Example commands for MPI run and outputs from running ncmpidump on the
! NC file produced by this example program:
!
! % mpif90 -O2 -o flexible_api flexible_api.f90 -lpnetcdf
! % mpiexec -n 4 ./flexible_api /pvfs2/wkliao/testfile.nc
!
! % ncmpidump /pvfs2/wkliao/testfile.nc
! netcdf testfile {
! // file format: CDF-5 (big variables)
! dimensions:
! x = 5 ;
! y = 16 ;
! variables:
! int var(y, x) ;
! data:
!
! var =
! 0, 0, 0, 0, 0,
! 0, 0, 0, 0, 0,
! 0, 0, 0, 0, 0,
! 0, 0, 0, 0, 0,
! 1, 1, 1, 1, 1,
! 1, 1, 1, 1, 1,
! 1, 1, 1, 1, 1,
! 1, 1, 1, 1, 1,
! 2, 2, 2, 2, 2,
! 2, 2, 2, 2, 2,
! 2, 2, 2, 2, 2,
! 2, 2, 2, 2, 2,
! 3, 3, 3, 3, 3,
! 3, 3, 3, 3, 3,
! 3, 3, 3, 3, 3,
! 3, 3, 3, 3, 3 ;
! }
!
subroutine check(err, message)
use mpi
use pnetcdf
implicit none
integer err
character(len=*) message
! It is a good idea to check returned value for possible error
if (err .NE. NF90_NOERR) then
write(6,*) trim(message), trim(nf90mpi_strerror(err))
call MPI_Abort(MPI_COMM_WORLD, -1, err)
end if
end subroutine check
program main
use mpi
use pnetcdf
implicit none
character(LEN=256) filename, cmd
integer err, nprocs, rank, i, j, ghost_len, ierr, get_args, dummy
integer cmode, ncid, varid, dimid(2)
integer(kind=MPI_OFFSET_KIND) nx, ny, global_nx, global_ny
integer(kind=MPI_OFFSET_KIND) starts(2), counts(2), nTypes
PARAMETER(nx=5, ny=4, ghost_len=3)
integer buf(nx+2*ghost_len, ny+2*ghost_len)
integer subarray
integer array_of_sizes(2), array_of_subsizes(2)
integer array_of_starts(2)
integer(kind=MPI_OFFSET_KIND) malloc_size, sum_size
logical verbose
call MPI_Init(err)
call MPI_Comm_rank(MPI_COMM_WORLD, rank, err)
call MPI_Comm_size(MPI_COMM_WORLD, nprocs, err)
! take filename from command-line argument if there is any
if (rank .EQ. 0) then
verbose = .TRUE.
filename = "testfile.nc"
ierr = get_args(2, cmd, filename, verbose, dummy)
endif
call MPI_Bcast(ierr, 1, MPI_INTEGER, 0, MPI_COMM_WORLD, err)
if (ierr .EQ. 0) goto 999
call MPI_Bcast(verbose, 1, MPI_LOGICAL, 0, MPI_COMM_WORLD, err)
call MPI_Bcast(filename, 256, MPI_CHARACTER, 0, MPI_COMM_WORLD, err)
! set parameters
global_nx = nx
global_ny = ny * nprocs
! first initialize the entire buffer to -1
buf = -1;
! assign values for non-ghost cells
do j=ghost_len+1, ny+ghost_len
do i=ghost_len+1, nx+ghost_len
buf(i, j) = rank
enddo
enddo
! define an MPI datatype using MPI_Type_create_subarray()
array_of_sizes(1) = nx + 2*ghost_len
array_of_sizes(2) = ny + 2*ghost_len
array_of_subsizes(1) = nx
array_of_subsizes(2) = ny
array_of_starts(1) = ghost_len ! MPI start index starts with 0
array_of_starts(2) = ghost_len
call MPI_Type_create_subarray(2, array_of_sizes, &
array_of_subsizes, array_of_starts, MPI_ORDER_FORTRAN, &
MPI_INTEGER, subarray, err)
call MPI_Type_commit(subarray, err)
! create file, truncate it if exists
cmode = IOR(NF90_CLOBBER, NF90_64BIT_DATA)
err = nf90mpi_create(MPI_COMM_WORLD, filename, cmode, &
MPI_INFO_NULL, ncid)
call check(err, 'In nf90mpi_create: ')
! define dimensions x and y
err = nf90mpi_def_dim(ncid, "y", global_ny, dimid(2))
call check(err, 'In nf90mpi_def_dim y: ')
err = nf90mpi_def_dim(ncid, "x", global_nx, dimid(1))
call check(err, 'In nf90mpi_def_dim x: ')
! define a 2D variable of integer type
err = nf90mpi_def_var(ncid, "var", NF90_INT, dimid, varid)
call check(err, 'In nf90mpi_def_var: ')
! do not forget to exit define mode
err = nf90mpi_enddef(ncid)
call check(err, 'In nf90mpi_enddef: ')
! now we are in data mode
! Note that in Fortran, array indices start with 1
starts(1) = 1
starts(2) = ny * rank + 1
counts(1) = nx
counts(2) = ny
nTypes = 1
! must explicitly use the argument keywords for the buffer type and
! number of buffer types, if any previous argument is skipped
err = nf90mpi_put_var_all(ncid, varid, buf, starts, counts, &
BUFCOUNT=nTypes, BUFTYPE=subarray)
call check(err, 'In nf90mpi_put_var_all: ')
call MPI_Type_free(subarray, err)
! close the file
err = nf90mpi_close(ncid)
call check(err, 'In nf90mpi_close: ')
! check if there is any PnetCDF internal malloc residue
998 format(A,I13,A)
err = nf90mpi_inq_malloc_size(malloc_size)
if (err == NF90_NOERR) then
call MPI_Reduce(malloc_size, sum_size, 1, MPI_INTEGER8, &
MPI_SUM, 0, MPI_COMM_WORLD, err)
if (rank .EQ. 0 .AND. sum_size .GT. 0_MPI_OFFSET_KIND) print 998, &
'heap memory allocated by PnetCDF internally has ', &
sum_size/1048576, ' MiB yet to be freed'
endif
999 call MPI_Finalize(err)
! call EXIT(0) ! EXIT() is a GNU extension
end program main
|