File: omp_init_lock.c

package info (click to toggle)
llvm-toolchain-11 1%3A11.0.1-2
  • links: PTS, VCS
  • area: main
  • in suites: bullseye
  • size: 995,808 kB
  • sloc: cpp: 4,767,656; ansic: 760,916; asm: 477,436; python: 170,940; objc: 69,804; lisp: 29,914; sh: 23,855; f90: 18,173; pascal: 7,551; perl: 7,471; ml: 5,603; awk: 3,489; makefile: 2,573; xml: 915; cs: 573; fortran: 503; javascript: 452
file content (42 lines) | stat: -rw-r--r-- 1,306 bytes parent folder | download | duplicates (22)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
// RUN: %libomp-compile-and-run
#include "omp_testsuite.h"
#include <stdio.h>

// This should be slightly less than KMP_I_LOCK_CHUNK, which is 1024
#define LOCKS_PER_ITER 1000
#define ITERATIONS (REPETITIONS + 1)

// This tests concurrently using locks on one thread while initializing new
// ones on another thread.  This exercises the global lock pool.
int test_omp_init_lock() {
  int i;
  omp_lock_t lcks[ITERATIONS * LOCKS_PER_ITER];
#pragma omp parallel for schedule(static) num_threads(NUM_TASKS)
  for (i = 0; i < ITERATIONS; i++) {
    int j;
    omp_lock_t *my_lcks = &lcks[i * LOCKS_PER_ITER];
    for (j = 0; j < LOCKS_PER_ITER; j++) {
      omp_init_lock(&my_lcks[j]);
    }
    for (j = 0; j < LOCKS_PER_ITER * 100; j++) {
      omp_set_lock(&my_lcks[j % LOCKS_PER_ITER]);
      omp_unset_lock(&my_lcks[j % LOCKS_PER_ITER]);
    }
  }
  // Wait until all repetitions are done.  The test is exercising growth of
  // the global lock pool, which does not shrink when no locks are allocated.
  {
    int j;
    for (j = 0; j < ITERATIONS * LOCKS_PER_ITER; j++) {
      omp_destroy_lock(&lcks[j]);
    }
  }

  return 0;
}

int main() {
  // No use repeating this test, since it's exercising a private global pool
  // which is not reset between test iterations.
  return test_omp_init_lock();
}