File: test_token_utils.py

package info (click to toggle)
python-friendly-traceback 0.7.62%2Bgit20240811.d7dbff6-1.1
  • links: PTS, VCS
  • area: main
  • in suites: forky, sid, trixie
  • size: 9,264 kB
  • sloc: python: 21,500; makefile: 4
file content (155 lines) | stat: -rw-r--r-- 3,492 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
from friendly_traceback import token_utils

# Note: most of the tests involving untokenize have
# been adapted from https://github.com/myint/untokenize


def check(source):
    tokens = token_utils.tokenize(source)
    new_source = token_utils.untokenize(tokens)
    assert source == new_source


def check_lines(source):
    lines = token_utils.get_lines(source)
    tokens = []
    for line in lines:
        tokens.extend(line)
    assert source == token_utils.untokenize(tokens)


def test_untokenize():
    check(
        '''

def zap():

    """Hello zap.

  """; 1


    x \t= \t\t  \t 1


'''
    )


def test_untokenize_with_tab_indentation():
    check(
        """
if True:
\tdef zap():
\t\tx \t= \t\t  \t 1
"""
    )


def test_untokenize_with_backslash_in_comment():
    check(
        r'''
def foo():
    """Hello foo."""
    def zap(): bar(1) # \
'''
    )


def test_untokenize_with_escaped_newline():
    check(
        r'''def foo():
    """Hello foo."""
    x = \
            1
'''
    )


def test_cpython_bug_35107():
    # Checking https://bugs.python.org/issue35107#msg328884
    check("#")
    check("#\n")


def test_last_line_empty():
    """If the last line contains only space characters with no newline
    Python's tokenizer drops this content. To ensure that the
    tokenize-untokenize returns the original value, we have introduced
    a fix in our utility functions"""

    source = "a\n  "
    source_2 = "a\n\t"
    check(source)
    check(source_2)

    check_lines(source)
    check_lines(source_2)


source1 = "a = b"
source2 = "a = b # comment\n"
source3 = """
if True:
    a = b # comment
"""
tokens1 = token_utils.tokenize(source1)
tokens2 = token_utils.tokenize(source2)
lines3 = token_utils.get_lines(source3)


def test_dedent():
    new_tokens = token_utils.dedent(lines3[2], 4)
    assert new_tokens == tokens2


def test_indent():
    new_tokens = token_utils.indent(tokens2, 4)
    new_line_a = token_utils.untokenize(new_tokens)
    new_line_b = token_utils.untokenize(lines3[2])
    assert new_line_a == new_line_b


def test_self():
    with open(__file__, "r") as f:
        source = f.read()
    check(source)


def test_find_substring_index():
    assert token_utils.find_substring_index(source2, source3) == -1
    assert token_utils.find_substring_index(source3, source2) == 3


def test_immediately_before_and_after():
    tokens = token_utils.get_significant_tokens("**/ =")
    assert tokens[0].immediately_before(tokens[1])
    assert tokens[1].immediately_after(tokens[0])
    assert not tokens[1].immediately_before(tokens[2])
    assert not tokens[2].immediately_after(tokens[1])

def test_unclosed_triple_quoted_string():
    with open("tests/unclosed.txt") as f:
        source = f.read()
        assert token_utils.untokenize(token_utils.tokenize(source)) == source

def test_unterminated_string():
    # See https://github.com/friendly-traceback/friendly-traceback/issues/241
    with open("tests/unterminated.txt") as f:
        source = f.read()
        assert token_utils.untokenize(token_utils.tokenize(source)) == source


def test_strip_commment():
    statement = "if True: # a comment"
    stripped = token_utils.strip_comment(statement)
    assert stripped.strip() == "if True:"


def test_invalid_octal():
    # See https://github.com/friendly-traceback/friendly-traceback/issues/242
    check("b = 0o1876 + 0o2")
    check("a = 0o23 + 0O2987")

def test_non_printable_character():
    check('print\x17("Hello")')