1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144
|
<!--
This file is part of Code_Saturne, a general-purpose CFD tool.
Copyright (C) 1998-2021 EDF S.A.
This program is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free Software
Foundation; either version 2 of the License, or (at your option) any later
version.
This program is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
details.
You should have received a copy of the GNU General Public License along with
this program; if not, write to the Free Software Foundation, Inc., 51 Franklin
Street, Fifth Floor, Boston, MA 02110-1301, USA.
-->
\page cs_ug_parallel Parallelism and periodicity
[TOC]
Parallelism basics
==================
Parallelism is based on domain partitioning: each processor is assigned
a part of the domain, and data for cells on parallel boundaries
is duplicated on neighboring processors in corresponding "ghost",
or "halo" cells (both terms are used interchangeably). Values in
these cells may be accessed just the same as values in regular cells.
Communication is only required when cell values are modified
using values from neighboring cells, as the values in the "halo" can
not be computed correctly (since the halo does not have access to all
its neighbors), so halo values must be updated by copying values from
the corresponding cells on the neighboring processor.
Compared to other tools using a similar system, a specificity of
code_saturne is the separation of the halo in two parts: a standard part,
containing cells shared through faces on parallel boundaries, and an
extended part, containing cells shared through vertices, which is
used mainly for least squares gradient reconstruction using an
extended neighborhood. Most updates need only to operate on the standard
halo, requiring less data communication than those on the extended halos.
\anchor fig_halo
\image html halo.svg "Parallel domain partitioning: halos"
Periodicity
-----------
Periodicity is handled using the same halo structures as parallelism,
with an additional treatment for vector and coordinate values: updating
coordinates requires applying the periodic transformation to the copied
values, and in the case of rotation, updating vector and tensor values
also requires applying the rotation transformation.
Ghost cells may be parallel, periodic, or both. The example of a pump
combining parallelism and periodicity is given in the following figure
In this example, all periodic boundaries match with boundaries on
the same domain, so halos are either parallel or periodic.
\anchor fig_parperio_pump
\image html rota_perio_parall.jpg "Combined parallelism and periodicity: halos" width=380px
Coding operations in parallel mode
----------------------------------
In parallel mode, the user must pay attention when performing
global operations. The following list is not exhaustive:
* calculation of extreme values on the domain (for instance, minimum
and maximum of some calculation values);
* test of the existence of a certain value (for instance, do faces
in a certain group exist?);
* verification of a condition on the domain (for instance, is a
given flow value reached somewhere?);
* counting out of entities (for instance, how many cells have
a nonzero source term value ?);
* global sum (for instance, calculation of a mass flow or the total
mass of a pollutant).
* Access to values in neighboring cells: with periodicity or parallelism,
some interior faces will be on parallel boundaries. When values
in these cells may have been modified locally, it is necessary to
"synchronize" the ghost values for that array, using functions such
as \ref synsca in Fortran or \ref cs_halo_sync_var in C, before
using the face neighbor values.
* In C code, the \ref cs_glob_rank_id and \ref cs_glob_n_ranks
global variables can be used to query the current rank id
(-1 in serial model, 0 to n-1 in parallel) and the number
of ranks.
- in Fortran, the matching variables are `irangp` and `nrangp`.
* The presence of periodicity is tested with the variable
<tt>cs_glob_mesh->n_init_perio</tt> in C, `iperio` in Fortran
(> 1 if periodicity is activated);
- The presence of rotation periodicity is tested with the
<tt>cs_glob_mesh->have_rotation_perio</tt> variable in C
(`iperot` in Fortran).
The user may refer to the different
[parallel operation](@ref cs_user_extra_operations_examples_parallel_operations_p)
examples present.
Care should be taken with the fact that the boundaries between
subdomains consist of **internal** faces shared between
two processors (these are indeed internal faces, even if they are
located at a "processor boundary". They should not be counted twice
(once per processor) during global operations using internal faces
(for instance, counting the internal faces per processor and
summing all the obtained numbers drives into over-evaluating the
number of internal faces of the initial mesh).
### Logging operations in parallel mode
When running in parallel, only the first rank actually produces outputs
when writing to `run_solver.log` using the `nfecra` logical unit
in Fortran, or \ref bft_printf or \ref cs_log_printf in C.
This avoids requiring tests in calling code, which would add clutter
an could easily be forgotten.
### File ouptut operations in parallel mode
When writing simple output to files, it is important to check for the
local rank, and to avoid writing to a same file from multiple processors,
unless dedicated features are used, such as the \ref cs_file.c functions.
### Some notes about periodicity
Note that periodic faces are not part of the domain boundary:
periodicity is interpreted as a "geometric" condition
rather than a classical boundary condition.
Some particular points should be reminded:
* Periodicity should work when the periodic boundaries are meshed
differently (periodicity of non-conforming faces), *except* for
the case of a 180 degree rotation periodicity with faces coupled
on the rotation axis.
* Rotation periodicity is incompatible with
* semi-transparent radiation,
* reinforced velocity-pressure coupling `ipucou=1`
* In the case of rotation periodicity with the uncoupled *R<sub>ij</sub> model,
some terms which are usually implicited are handled explicitely.
|