1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327
|
#! /usr/bin/env python3
# $Id: test_io.py 9994 2024-12-04 13:32:39Z milde $
# Author: Lea Wiemann <LeWiemann@gmail.com>
# Copyright: This module has been placed in the public domain.
"""
Test module for `docutils.io`.
"""
import codecs
import locale
import sys
import unittest
import warnings
from io import StringIO, BytesIO
from pathlib import Path
if __name__ == '__main__':
# prepend the "docutils root" to the Python library path
# so we import the local `docutils` package.
sys.path.insert(0, str(Path(__file__).resolve().parents[1]))
from docutils import io as du_io
# For when we intentionally do things that trigger EncodingWarning
# (When using ``-X warn_default_encoding`` or PYTHONWARNDEFAULTENCODING=1)
# See: https://docs.python.org/3/library/io.html#io-encoding-warning
if sys.version_info[:2] > (3, 9):
SUPPRESS_ENCODING_WARNING = sys.flags.warn_default_encoding
else:
SUPPRESS_ENCODING_WARNING = False
EncodingWarning = UnicodeWarning # NoQA: A001 (builtin in Py > 0.9)
# DATA_ROOT is ./test/data/ from the docutils root
DATA_ROOT = Path(__file__).parent / 'data'
# normalize the preferred encoding's name:
with warnings.catch_warnings():
if SUPPRESS_ENCODING_WARNING:
warnings.filterwarnings('ignore', category=EncodingWarning)
preferredencoding = codecs.lookup(
locale.getpreferredencoding(do_setlocale=False)).name
# Stub: Buffer with 'strict' auto-conversion of input to byte string:
class BBuf(BytesIO):
def write(self, data):
if isinstance(data, str):
data.encode('ascii', 'strict')
super().write(data)
# Stub: Buffer expecting unicode string:
class UBuf(StringIO):
def write(self, data):
# emulate Python 3 handling of stdout, stderr
if isinstance(data, bytes):
raise TypeError('must be unicode, not bytes')
super().write(data)
class mock_stdout(UBuf):
encoding = 'utf-8'
def __init__(self):
self.buffer = BBuf()
super().__init__()
class HelperTests(unittest.TestCase):
def test_check_encoding_true(self):
"""Return `True` if lookup returns the same codec"""
self.assertEqual(True, du_io.check_encoding(mock_stdout, 'utf-8'))
self.assertEqual(True, du_io.check_encoding(mock_stdout, 'utf_8'))
self.assertEqual(True, du_io.check_encoding(mock_stdout, 'utf8'))
self.assertEqual(True, du_io.check_encoding(mock_stdout, 'UTF-8'))
def test_check_encoding_false(self):
"""Return `False` if lookup returns different codecs"""
self.assertEqual(False, du_io.check_encoding(mock_stdout, 'ascii'))
self.assertEqual(False, du_io.check_encoding(mock_stdout, 'latin-1'))
def test_check_encoding_none(self):
"""Cases where the comparison fails."""
# stream.encoding is None:
self.assertEqual(None,
du_io.check_encoding(du_io.FileInput(encoding=None),
'ascii'))
# stream.encoding does not exist:
self.assertEqual(None, du_io.check_encoding(BBuf, 'ascii'))
# encoding is None or empty string:
self.assertEqual(None, du_io.check_encoding(mock_stdout, None))
self.assertEqual(None, du_io.check_encoding(mock_stdout, ''))
# encoding is invalid
self.assertEqual(None, du_io.check_encoding(mock_stdout, 'UTF-9'))
def test_error_string(self):
us = '\xfc' # bytes(us) fails
bs = b'\xc3\xbc' # str(bs) returns repr(bs)
self.assertEqual('Exception: spam',
du_io.error_string(Exception('spam')))
self.assertEqual('IndexError: ' + str(bs),
du_io.error_string(IndexError(bs)))
self.assertEqual('ImportError: %s' % us,
du_io.error_string(ImportError(us)))
class InputTests(unittest.TestCase):
def test_bom_handling(self):
# Provisional:
# default input encoding will change to UTF-8 in Docutils 0.22
source = '\ufeffdata\n\ufeff blah\n'
expected = 'data\n\ufeff blah\n' # only leading ZWNBSP removed
input_ = du_io.StringInput(source=source.encode('utf-16-be'),
encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
self.assertEqual(expected, input_.read())
input_ = du_io.StringInput(source=source.encode('utf-16-le'),
encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
self.assertEqual(expected, input_.read())
input_ = du_io.StringInput(source=source.encode('utf-8'),
encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
self.assertEqual(expected, input_.read())
# With `str` input all ZWNBSPs are still there.
input_ = du_io.StringInput(source=source)
self.assertEqual(source, input_.read())
def test_encoding_declaration(self):
input_ = du_io.StringInput(source=b"""\
.. -*- coding: ascii -*-
data
blah
""", encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
data = input_.read() # noqa: F841
self.assertEqual('ascii', input_.successful_encoding)
input_ = du_io.StringInput(source=b"""\
#! python
# -*- coding: ascii -*-
print("hello world")
""", encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
data = input_.read() # noqa: F841
self.assertEqual('ascii', input_.successful_encoding)
input_ = du_io.StringInput(source=b"""\
#! python
# extraneous comment; prevents coding slug from being read
# -*- coding: ascii -*-
print("hello world")
""")
self.assertNotEqual(input_.successful_encoding, 'ascii')
def test_decode_unicode(self):
# With the special value "unicode" or "Unicode":
uniinput = du_io.Input(encoding='unicode')
# keep unicode instances as-is
self.assertEqual('ja', uniinput.decode('ja'))
# raise AssertionError if data is not a `str` instance
with self.assertRaises(AssertionError):
uniinput.decode(b'ja')
class OutputTests(unittest.TestCase):
bdata = b'\xfc'
udata = '\xfc'
def setUp(self):
self.bdrain = BBuf()
"""Buffer accepting binary strings (bytes)"""
self.udrain = UBuf()
"""Buffer accepting unicode strings"""
self.mock_stdout = mock_stdout()
"""Stub of sys.stdout under Python 3"""
def test_write_unicode(self):
fo = du_io.FileOutput(destination=self.udrain, encoding='unicode',
autoclose=False)
fo.write(self.udata)
self.assertEqual(self.udata, self.udrain.getvalue())
def test_write_utf8(self):
fo = du_io.FileOutput(destination=self.udrain, encoding='utf-8',
autoclose=False)
fo.write(self.udata)
self.assertEqual(self.udata, self.udrain.getvalue())
def test_FileOutput_hande_io_errors_deprection_warning(self):
with self.assertWarnsRegex(DeprecationWarning,
'"handle_io_errors" is ignored'):
du_io.FileOutput(handle_io_errors=True)
# With destination in binary mode, data must be binary string
# and is written as-is:
def test_write_bytes(self):
fo = du_io.FileOutput(destination=self.bdrain, encoding='utf-8',
mode='wb', autoclose=False)
fo.write(self.bdata)
self.assertEqual(self.bdata, self.bdrain.getvalue())
def test_write_bytes_to_stdout(self):
# try writing data to `destination.buffer`, if data is
# instance of `bytes` and writing to `destination` fails:
fo = du_io.FileOutput(destination=self.mock_stdout)
fo.write(self.bdata)
self.assertEqual(self.bdata,
self.mock_stdout.buffer.getvalue())
def test_encoding_clash_resolved(self):
fo = du_io.FileOutput(destination=self.mock_stdout,
encoding='latin1', autoclose=False)
fo.write(self.udata)
self.assertEqual(self.udata.encode('latin1'),
self.mock_stdout.buffer.getvalue())
def test_encoding_clash_nonresolvable(self):
del self.mock_stdout.buffer
fo = du_io.FileOutput(destination=self.mock_stdout,
encoding='latin1', autoclose=False)
self.assertRaises(ValueError, fo.write, self.udata)
class ErrorOutputTests(unittest.TestCase):
def test_defaults(self):
e = du_io.ErrorOutput()
self.assertEqual(sys.stderr, e.destination)
def test_bbuf(self):
buf = BBuf() # buffer storing byte string
e = du_io.ErrorOutput(buf, encoding='ascii')
# write byte-string as-is
e.write(b'b\xfc')
self.assertEqual(b'b\xfc', buf.getvalue())
# encode unicode data with backslashescape fallback replacement:
e.write(' u\xfc')
self.assertEqual(b'b\xfc u\\xfc', buf.getvalue())
# handle Exceptions with Unicode string args
# unicode(Exception('e\xfc')) # fails in Python < 2.6
e.write(AttributeError(' e\xfc'))
self.assertEqual(b'b\xfc u\\xfc e\\xfc', buf.getvalue())
# encode with `encoding` attribute
e.encoding = 'utf-8'
e.write(' u\xfc')
self.assertEqual(b'b\xfc u\\xfc e\\xfc u\xc3\xbc', buf.getvalue())
def test_ubuf(self):
buf = UBuf() # buffer only accepting unicode string
# decode of binary strings
e = du_io.ErrorOutput(buf, encoding='ascii')
e.write(b'b\xfc')
# use REPLACEMENT CHARACTER
self.assertEqual(buf.getvalue(), 'b\ufffd')
# write Unicode string and Exceptions with Unicode args
e.write(' u\xfc')
self.assertEqual(buf.getvalue(), 'b\ufffd u\xfc')
e.write(AttributeError(' e\xfc'))
self.assertEqual(buf.getvalue(), 'b\ufffd u\xfc e\xfc')
# decode with `encoding` attribute
e.encoding = 'latin1'
e.write(b' b\xfc')
self.assertEqual(buf.getvalue(), 'b\ufffd u\xfc e\xfc b\xfc')
class FileInputTests(unittest.TestCase):
# test input encoding auto-detection:
#
# Up to Docutils 0.18, auto-detection was not used under Python 3
# unless reading a file with Python's default encoding failed
def test_bom_utf_8(self):
"""Drop optional BOM from utf-8 encoded files.
"""
with warnings.catch_warnings():
if SUPPRESS_ENCODING_WARNING:
warnings.filterwarnings('ignore', category=EncodingWarning)
source = du_io.FileInput(source_path=DATA_ROOT/'utf-8-sig.rst',
encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
self.assertTrue(source.read().startswith('Grüße'))
def test_bom_utf_16(self):
"""Drop BOM from utf-16 encoded files, use correct encoding.
"""
# Assert correct decoding, BOM is gone.
with warnings.catch_warnings():
if SUPPRESS_ENCODING_WARNING:
warnings.filterwarnings('ignore', category=EncodingWarning)
source = du_io.FileInput(source_path=DATA_ROOT/'utf-16-le-sig.rst',
encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
self.assertTrue(source.read().startswith('Grüße'))
def test_coding_slug(self):
"""Use self-declared encoding.
"""
with warnings.catch_warnings():
if SUPPRESS_ENCODING_WARNING:
warnings.filterwarnings('ignore', category=EncodingWarning)
source = du_io.FileInput(source_path=DATA_ROOT/'latin2.rst',
encoding=None)
with self.assertWarnsRegex(DeprecationWarning, 'auto-detection'):
self.assertTrue(source.read().endswith('škoda\n'))
def test_fallback_utf8(self):
"""Try 'utf-8', if encoding is not specified in the source."""
with warnings.catch_warnings():
if SUPPRESS_ENCODING_WARNING:
warnings.filterwarnings('ignore', category=EncodingWarning)
source = du_io.FileInput(source_path=DATA_ROOT/'utf8.rst',
encoding=None)
self.assertEqual('Grüße\n', source.read())
def test_readlines(self):
source = du_io.FileInput(source_path=DATA_ROOT/'include.rst')
data = source.readlines()
self.assertEqual(['Some include text.\n'], data)
if __name__ == '__main__':
unittest.main()
|