File: README.md

package info (click to toggle)
llvm-toolchain-15 1%3A15.0.6-4
  • links: PTS, VCS
  • area: main
  • in suites: bookworm
  • size: 1,554,644 kB
  • sloc: cpp: 5,922,452; ansic: 1,012,136; asm: 674,362; python: 191,568; objc: 73,855; f90: 42,327; lisp: 31,913; pascal: 11,973; javascript: 10,144; sh: 9,421; perl: 7,447; ml: 5,527; awk: 3,523; makefile: 2,520; xml: 885; cs: 573; fortran: 567
file content (34 lines) | stat: -rw-r--r-- 1,425 bytes parent folder | download | duplicates (15)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
# Size distributions for memory functions under specific workloads

This folder contains a set of files that are included from `libc/benchmarks/MemorySizeDistributions.cpp`.

Offloading this data to individual files helps
 - C++ editors (large arrays are usually not well handled by editors),
 - and allows processing data by other tools to perform analysis or graph rendering.

 ## Format

Most filenames are of the form `{MemoryFunctionName}{WorkloadID}.csv`. They contain a single line of comma separated real values representing the probability that a particular size occurs. e.g.
 - `"0,1"` indicates that only the size `1` occurs,
 - `"0.5,0.5"` indicates sizes `0` and `1` occur with the same frequency.

 These files usually contains sizes from `0` to `4096` inclusive. To save on space trailing zeros are discarded.

 ## Workloads

As identified in the [automemcpy](https://research.google/pubs/pub50338/) paper:
  - `GoogleA` <-> `service 4`
  - `GoogleB` <-> `database 1`
  - `GoogleD` <-> `storage`
  - `GoogleL` <-> `logging`
  - `GoogleM` <-> `service 2`
  - `GoogleQ` <-> `database 2`
  - `GoogleS` <-> `database 3`
  - `GoogleU` <-> `service 3`
  - `GoogleW` <-> `service 1`

`Uniform384To4096` is an additional synthetic workload that simply returns a uniform repartition of the sizes from `384` to `4096` inclusive.

## Note

Except for `GoogleD`, all distributions are gathered over one week worth of data.