File: mtest.c

package info (click to toggle)
parmetis 4.0.3-5
  • links: PTS, VCS
  • area: non-free
  • in suites: bullseye, buster, sid
  • size: 25,384 kB
  • ctags: 3,256
  • sloc: ansic: 41,872; makefile: 298; sh: 190; perl: 25
file content (100 lines) | stat: -rw-r--r-- 2,701 bytes parent folder | download | duplicates (3)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
/*
 * Copyright 1997, Regents of the University of Minnesota
 *
 * main.c
 * 
 * This file contains code for testing teh adaptive partitioning routines
 *
 * Started 5/19/97
 * George
 *
 * $Id: mtest.c,v 1.3 2003/07/25 14:31:47 karypis Exp $
 *
 */

#include <parmetisbin.h>


/*************************************************************************
* Let the game begin
**************************************************************************/
int main(int argc, char *argv[])
{
  idx_t i, mype, npes, nelms;
  idx_t *part, *eptr;
  mesh_t mesh;
  MPI_Comm comm;
  idx_t wgtflag, numflag, edgecut, nparts, options[10];
  idx_t mgcnum = -1, mgcnums[5] = {-1, 2, 3, 4, 2}, esizes[5] = {-1, 3, 4, 8, 4};
  real_t *tpwgts, ubvec[MAXNCON];

  gk_malloc_init();

  MPI_Init(&argc, &argv);
  MPI_Comm_dup(MPI_COMM_WORLD, &comm);
  gkMPI_Comm_size(comm, &npes);
  gkMPI_Comm_rank(comm, &mype);

  if (argc < 2) {
    if (mype == 0)
      printf("Usage: %s <mesh-file> [NCommonNodes]\n", argv[0]);

    MPI_Finalize();
    exit(0);
  }

  ParallelReadMesh(&mesh, argv[1], comm); 
  mgcnum = mgcnums[mesh.etype];
  mesh.ncon = 1;

  if (argc > 2)
    mgcnum = atoi(argv[2]);

  if (mype == 0) printf("MGCNUM: %"PRIDX"\n", mgcnum);

  nparts = npes;
  tpwgts = rmalloc(nparts*mesh.ncon, "tpwgts");
  for (i=0; i<nparts*mesh.ncon; i++)
    tpwgts[i] = 1.0/(real_t)(nparts);

  for (i=0; i<mesh.ncon; i++)
    ubvec[i] = UNBALANCE_FRACTION;

  part = imalloc(mesh.nelms, "part");

  numflag = wgtflag = 0;
  options[0] = 1;
  options[PMV3_OPTION_DBGLVL] = 7;
  options[PMV3_OPTION_SEED] = 0;

  nelms = mesh.elmdist[mype+1]-mesh.elmdist[mype];
  eptr = ismalloc(nelms+1, esizes[mesh.etype], "main; eptr");
  MAKECSR(i, nelms, eptr);
  eptr[nelms]--; /* make the last element different */
  ParMETIS_V3_PartMeshKway(mesh.elmdist, eptr, mesh.elements, NULL, &wgtflag, 
              &numflag, &(mesh.ncon), &mgcnum, &nparts, tpwgts, ubvec, options, 
	      &edgecut, part, &comm);
 
/* 
  graph = ParallelMesh2Dual(&mesh, mgcnum, comm);
  MPI_Barrier(comm);

  MPI_Allreduce((void *)&(graph->nedges), (void *)&gnedges, 1, IDX_T, MPI_SUM, comm);
  if (mype == 0)
    printf("Completed Dual Graph -- Nvtxs: %"PRIDX", Nedges: %"PRIDX"\n", graph->gnvtxs, gnedges/2);

  numflag = wgtflag = 0;
  ParMETIS_V3_PartKway(graph->vtxdist, graph->xadj, graph->adjncy, NULL, NULL, &wgtflag,
  &numflag, &(graph->ncon), &nparts, tpwgts, ubvec, options, &edgecut, part, &comm);
  gk_free((void **)&(graph.vtxdist), &(graph.xadj), &(graph.vwgt), &(graph.adjncy), &(graph.adjwgt), LTERM);
*/ 

  gk_free((void **)&part, &tpwgts, &eptr, LTERM);
  MPI_Comm_free(&comm);
  MPI_Finalize();

  gk_malloc_cleanup(0);
  return 0;
}