1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295
|
"""
This is a sphinx extension to freeze your broken reference problems
when using ``nitpicky = True``.
The basic operation is:
1. Add this extension to your ``conf.py`` extensions.
2. Add ``missing_references_write_json = True`` to your ``conf.py``
3. Run sphinx-build. It will generate ``missing-references.json``
next to your ``conf.py``.
4. Remove ``missing_references_write_json = True`` from your
``conf.py`` (or set it to ``False``)
5. Run sphinx-build again, and ``nitpick_ignore`` will
contain all of the previously failed references.
"""
from collections import defaultdict
import json
import logging
from pathlib import Path
from docutils.utils import get_source_line
from docutils import nodes
from sphinx.util import logging as sphinx_logging
import matplotlib
logger = sphinx_logging.getLogger(__name__)
class MissingReferenceFilter(logging.Filter):
"""
A logging filter designed to record missing reference warning messages
for use by this extension
"""
def __init__(self, app):
self.app = app
super().__init__()
def _record_reference(self, record):
if not (getattr(record, 'type', '') == 'ref' and
isinstance(getattr(record, 'location', None), nodes.Node)):
return
if not hasattr(self.app.env, "missing_references_warnings"):
self.app.env.missing_references_warnings = defaultdict(set)
record_missing_reference(self.app,
self.app.env.missing_references_warnings,
record.location)
def filter(self, record):
self._record_reference(record)
return True
def record_missing_reference(app, record, node):
domain = node["refdomain"]
typ = node["reftype"]
target = node["reftarget"]
location = get_location(node, app)
domain_type = "{}:{}".format(domain, typ)
record[(domain_type, target)].add(location)
def record_missing_reference_handler(app, env, node, contnode):
"""
When the sphinx app notices a missing reference, it emits an
event which calls this function. This function records the missing
references for analysis at the end of the sphinx build.
"""
if not app.config.missing_references_enabled:
# no-op when we are disabled.
return
if not hasattr(env, "missing_references_events"):
env.missing_references_events = defaultdict(set)
record_missing_reference(app, env.missing_references_events, node)
def get_location(node, app):
"""
Given a docutils node and a sphinx application, return a string
representation of the source location of this node.
Usually, this will be of the form "path/to/file:linenumber". Two
special values can be emitted, "<external>" for paths which are
not contained in this source tree (e.g. docstrings included from
other modules) or "<unknown>", inidcating that the sphinx application
cannot locate the original source file (usually because an extension
has injected text into the sphinx parsing engine).
"""
source, line = get_source_line(node)
if source:
# 'source' can have the form '/some/path:docstring of some.api' but the
# colons are forbidden on windows, but on posix just passes through.
path, *post = source.partition(':')
post = ''.join(post)
# We locate references relative to the parent of the doc
# directory, which for matplotlib, will be the root of the
# matplotlib repo. When matplotlib is not an editable install
# weird things will happen, but we can't totally recover from
# that.
basepath = Path(app.srcdir).parent.resolve()
fullpath = Path(path).resolve()
try:
path = fullpath.relative_to(basepath)
except ValueError:
# Sometimes docs directly contain e.g. docstrings
# from installed modules, and we record those as
# <external> so as to be independent of where the
# module was installed
path = Path("<external>") / fullpath.name
# Ensure that all reported paths are POSIX so that docs
# on windows result in the same warnings in the JSON file.
path = path.as_posix()
else:
path = "<unknown>"
post = ''
if not line:
line = ""
return f"{path}{post}:{line}"
def _truncate_location(location):
"""
Cuts off anything after the first colon in location strings.
This allows for easy comparison even when line numbers chagne
(as they do regularly).
"""
return location.split(":", 1)[0]
def _warn_unused_missing_references(app):
if not app.config.missing_references_warn_unused_ignores:
return
# We can only warn if we are building from a source install
# otherwise, we just have to skip this step.
basepath = Path(matplotlib.__file__).parent.parent.parent.resolve()
srcpath = Path(app.srcdir).parent.resolve()
if basepath != srcpath:
return
# This is a dictionary of {(domain_type, target): locations}
references_ignored = getattr(
app.env, 'missing_references_ignored_references', {})
references_events = getattr(app.env, 'missing_references_events', {})
# Warn about any reference which is no longer missing.
for (domain_type, target), locations in references_ignored.items():
missing_reference_locations = [
_truncate_location(location)
for location in references_events.get((domain_type, target), [])]
# For each ignored reference location, ensure a missing reference
# was observed. If it wasn't observed, issue a warning.
for ignored_reference_location in locations:
short_location = _truncate_location(ignored_reference_location)
if short_location not in missing_reference_locations:
msg = (f"Reference {domain_type} {target} for "
f"{ignored_reference_location} can be removed"
f" from {app.config.missing_references_filename}."
"It is no longer a missing reference in the docs.")
logger.warning(msg,
location=ignored_reference_location,
type='ref',
subtype=domain_type)
def save_missing_references_handler(app, exc):
"""
At the end of the sphinx build, check that all lines of the existing JSON
file are still necessary.
If the configuration value ``missing_references_write_json`` is set
then write a new JSON file containing missing references.
"""
if not app.config.missing_references_enabled:
# no-op when we are disabled.
return
_warn_unused_missing_references(app)
json_path = (Path(app.confdir) /
app.config.missing_references_filename)
references_warnings = getattr(app.env, 'missing_references_warnings', {})
if app.config.missing_references_write_json:
_write_missing_references_json(references_warnings, json_path)
def _write_missing_references_json(records, json_path):
"""
Convert ignored references to a format which we can write as JSON
Convert from ``{(domain_type, target): locations}`` to
``{domain_type: {target: locations}}`` since JSON can't serialize tuples.
"""
# Sorting records and keys avoids needlessly big diffs when
# missing_references.json is regenerated.
transformed_records = defaultdict(dict)
for (domain_type, target), paths in records.items():
transformed_records[domain_type][target] = sorted(paths)
with json_path.open("w") as stream:
json.dump(transformed_records, stream, sort_keys=True, indent=2)
def _read_missing_references_json(json_path):
"""
Convert from the JSON file to the form used internally by this
extension.
The JSON file is stored as ``{domain_type: {target: [locations,]}}``
since JSON can't store dictionary keys which are tuples. We convert
this back to ``{(domain_type, target):[locations]}`` for internal use.
"""
with json_path.open("r") as stream:
data = json.load(stream)
ignored_references = {}
for domain_type, targets in data.items():
for target, locations in targets.items():
ignored_references[(domain_type, target)] = locations
return ignored_references
def prepare_missing_references_handler(app):
"""
Handler called to initialize this extension once the configuration
is ready.
Reads the missing references file and populates ``nitpick_ignore`` if
appropriate.
"""
if not app.config.missing_references_enabled:
# no-op when we are disabled.
return
sphinx_logger = logging.getLogger('sphinx')
missing_reference_filter = MissingReferenceFilter(app)
for handler in sphinx_logger.handlers:
if (isinstance(handler, sphinx_logging.WarningStreamHandler)
and missing_reference_filter not in handler.filters):
# This *must* be the first filter, because subsequent filters
# throw away the node information and then we can't identify
# the reference uniquely.
handler.filters.insert(0, missing_reference_filter)
app.env.missing_references_ignored_references = {}
json_path = (Path(app.confdir) /
app.config.missing_references_filename)
if not json_path.exists():
return
ignored_references = _read_missing_references_json(json_path)
app.env.missing_references_ignored_references = ignored_references
# If we are going to re-write the JSON file, then don't suppress missing
# reference warnings. We want to record a full list of missing references
# for use later. Otherwise, add all known missing references to
# ``nitpick_ignore```
if not app.config.missing_references_write_json:
app.config.nitpick_ignore.extend(ignored_references.keys())
def setup(app):
app.add_config_value("missing_references_enabled", True, "env")
app.add_config_value("missing_references_write_json", False, "env")
app.add_config_value("missing_references_warn_unused_ignores", True, "env")
app.add_config_value("missing_references_filename",
"missing-references.json", "env")
app.connect("builder-inited", prepare_missing_references_handler)
app.connect("missing-reference", record_missing_reference_handler)
app.connect("build-finished", save_missing_references_handler)
return {'parallel_read_safe': True}
|