File: conda-repodata.rst

package info (click to toggle)
python-msgspec 0.19.0-2
  • links: PTS, VCS
  • area: main
  • in suites: forky, trixie
  • size: 6,356 kB
  • sloc: javascript: 23,944; ansic: 20,540; python: 20,465; makefile: 29; sh: 19
file content (118 lines) | stat: -rw-r--r-- 4,550 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
Conda Repodata
==============

This example benchmarks using different JSON libraries to parse and query the
`current_repodata.json`_ file from conda-forge_. This is a medium-sized (~14
MiB) JSON file containing nested metadata about every package on conda-forge.

The following libraries are compared:

- json_
- ujson_
- orjson_
- simdjson_
- msgspec_

This benchmark measures how long it takes each library to decode the
``current_repodata.json`` file, extract the name and size of each package, and
determine the top 10 packages by file size.

**Results**

.. raw:: html

    <div id="bench-conda-repodata"></div>

.. code-block:: text

   $ python query_repodata.py
   json: 139.14 ms
   ujson: 124.91 ms
   orjson: 91.69 ms
   simdjson: 66.40 ms
   msgspec: 25.73 ms


**Commentary**

- All of these are fairly quick, library choice likely doesn't matter at all
  for simple scripts on small- to medium-sized data.

- While ``orjson`` is faster than ``json``, the difference between them is only
  ~30%. Creating python objects dominates the execution time of any well
  optimized decoding library. How fast the underlying JSON parser is matters,
  but JSON optimizations can only get you so far if you're still creating a new
  Python object for every node in the JSON object.

- ``simdjson`` is much more performant. This is partly due to the SIMD
  optimizations it uses, but mostly it's due to not creating so many Python
  objects. ``simdjson`` first parses a JSON blob into a proxy object. It then
  lazily creates Python objects as needed as different fields are accessed.
  This means you only pay the cost of creating Python objects for the fields
  you use; a query that only accesses a few fields runs much faster since not
  as many Python objects are created. The downside is every attribute access
  results in some indirection as new objects are created

- ``msgspec`` is the fastest option tested. It relies on defining a known
  schema beforehand. We don't define the schema for the entire structure, only
  for the fields we access. Only fields that are part of the schema are
  decoded, with a new Python object created for each. This allocates the same
  number of objects as ``simdjson``, but does it all at once, avoiding
  indirection costs later on during use. See :ref:`this performance tip
  <avoid-decoding-unused-fields>` for more information.

**Source**

The full example source can be found `here
<https://github.com/jcrist/msgspec/blob/main/examples/conda-repodata>`__.

.. literalinclude:: ../../../examples/conda-repodata/query_repodata.py
    :language: python

.. raw:: html

    +    <!-- needs package libjs-vega -->
    <script src="file:///usr/share/javascript/vega/vega.min.js"></script>
    <script src="file:///usr/share/doc/python-msgspec-doc/html/vega-lite.min.js"></script>
    <script src="file:///usr/share/doc/python-msgspec-doc/html/vega-embed.min.js"></script>
    <script type="text/javascript">
    var spec = {
        "$schema": "https://vega.github.io/schema/vega-lite/v5.json",
        "title": {
            "text": "Benchmark - Processing current_repodata.json",
            "anchor": "start"
        },
        "config": {"view": {"discreteWidth": 400, "stroke": null}},
        "data": {
            "values": [
            {"library": "json", "time": 139.14},
            {"library": "ujson", "time": 124.91},
            {"library": "orjson", "time": 91.69},
            {"library": "simdjson", "time": 66.40},
            {"library": "msgspec", "time": 25.73}
            ]
        },
        "transform": [
            {
                "calculate": `join([format(datum.time, '.3'), 'ms'], ' ')`,
                "as": "tooltip",
            }
        ],
        "mark": "bar",
        "encoding": {
            "tooltip": {"field": "tooltip", "type": "nominal"},
            "x": {"field": "library", "type": "nominal", "axis": {"labelAngle": 0, "title": null}, "sort": {"field": "time", "order": "descending"}},
            "y": {"field": "time", "type": "quantitative", "axis": {"title": "time (ms)", "grid": false}}
        }
    };
    vegaEmbed("#bench-conda-repodata", spec);
    </script>


.. _conda-forge: https://conda-forge.org/
.. _current_repodata.json: https://conda.anaconda.org/conda-forge/noarch/current_repodata.json
.. _json: https://docs.python.org/3/library/json.html
.. _ujson: https://github.com/ultrajson/ultrajson
.. _msgspec: https://jcristharif.com/msgspec/
.. _orjson: https://github.com/ijl/orjson
.. _simdjson: https://github.com/TkTech/pysimdjson