1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926
|
"""Docopt is a Pythonic command-line interface parser that will make you smile.
Now: with spellcheck, flag extension (de-abbreviation), and capitalization fixes.
(but only when unambiguous)
* Licensed under terms of MIT license (see LICENSE-MIT)
Contributors (roughly in chronological order):
* Copyright (c) 2012 Andrew Kassen <atkassen@ucdavis.edu>
* Copyright (c) 2012 jeffrimko <jeffrimko@gmail.com>
* Copyright (c) 2012 Andrew Sutton <met48@met48.com>
* Copyright (c) 2012 Andrew Sutton <met48@met48.com>
* Copyright (c) 2012 Nima Johari <nimajohari@gmail.com>
* Copyright (c) 2012-2013 Vladimir Keleshev, vladimir@keleshev.com
* Copyright (c) 2014-2018 Matt Boersma <matt@sprout.org>
* Copyright (c) 2016 amir <ladsgroup@gmail.com>
* Copyright (c) 2015 Benjamin Bach <benjaoming@gmail.com>
* Copyright (c) 2017 Oleg Bulkin <o.bulkin@gmail.com>
* Copyright (c) 2018 Iain Barnett <iainspeed@gmail.com>
* Copyright (c) 2019 itdaniher, itdaniher@gmail.com
"""
from __future__ import annotations
import re
import sys
from typing import Any
from typing import Callable
from typing import NamedTuple
from typing import Tuple
from typing import Type
from typing import Union
from typing import cast
from ._version import __version__ as __version__
__all__ = ["docopt", "DocoptExit"]
def levenshtein_norm(source: str, target: str) -> float:
"""Calculates the normalized Levenshtein distance between two string
arguments. The result will be a float in the range [0.0, 1.0], with 1.0
signifying the biggest possible distance between strings with these lengths
"""
# Compute Levenshtein distance using helper function. The max is always
# just the length of the longer string, so this is used to normalize result
# before returning it
distance = levenshtein(source, target)
return float(distance) / max(len(source), len(target))
def levenshtein(source: str, target: str) -> int:
"""Computes the Levenshtein
(https://en.wikipedia.org/wiki/Levenshtein_distance)
and restricted Damerau-Levenshtein
(https://en.wikipedia.org/wiki/Damerau%E2%80%93Levenshtein_distance)
distances between two Unicode strings with given lengths using the
Wagner-Fischer algorithm
(https://en.wikipedia.org/wiki/Wagner%E2%80%93Fischer_algorithm).
These distances are defined recursively, since the distance between two
strings is just the cost of adjusting the last one or two characters plus
the distance between the prefixes that exclude these characters (e.g. the
distance between "tester" and "tested" is 1 + the distance between "teste"
and "teste"). The Wagner-Fischer algorithm retains this idea but eliminates
redundant computations by storing the distances between various prefixes in
a matrix that is filled in iteratively.
"""
# Create matrix of correct size (this is s_len + 1 * t_len + 1 so that the
# empty prefixes "" can also be included). The leftmost column represents
# transforming various source prefixes into an empty string, which can
# always be done by deleting all characters in the respective prefix, and
# the top row represents transforming the empty string into various target
# prefixes, which can always be done by inserting every character in the
# respective prefix. The ternary used to build the list should ensure that
# this row and column are now filled correctly
s_range = range(len(source) + 1)
t_range = range(len(target) + 1)
matrix = [[(i if j == 0 else j) for j in t_range] for i in s_range]
# Iterate through rest of matrix, filling it in with Levenshtein
# distances for the remaining prefix combinations
for i in s_range[1:]:
for j in t_range[1:]:
# Applies the recursive logic outlined above using the values
# stored in the matrix so far. The options for the last pair of
# characters are deletion, insertion, and substitution, which
# amount to dropping the source character, the target character,
# or both and then calculating the distance for the resulting
# prefix combo. If the characters at this point are the same, the
# situation can be thought of as a free substitution
del_dist = matrix[i - 1][j] + 1
ins_dist = matrix[i][j - 1] + 1
sub_trans_cost = 0 if source[i - 1] == target[j - 1] else 1
sub_dist = matrix[i - 1][j - 1] + sub_trans_cost
# Choose option that produces smallest distance
matrix[i][j] = min(del_dist, ins_dist, sub_dist)
# At this point, the matrix is full, and the biggest prefixes are just the
# strings themselves, so this is the desired distance
return matrix[len(source)][len(target)]
class DocoptLanguageError(Exception):
"""Error in construction of usage-message by developer."""
class DocoptExit(SystemExit):
"""Exit in case user invoked program with incorrect arguments."""
usage = ""
def __init__(
self,
message: str = "",
collected: list[Pattern] | None = None,
left: list[Pattern] | None = None,
) -> None:
self.collected = collected if collected is not None else []
self.left = left if left is not None else []
SystemExit.__init__(self, (message + "\n" + self.usage).strip())
class Pattern:
def __init__(
self, name: str | None, value: list[str] | str | int | None = None
) -> None:
self._name, self.value = name, value
@property
def name(self) -> str | None:
return self._name
def __eq__(self, other) -> bool:
return repr(self) == repr(other)
def __hash__(self) -> int:
return hash(repr(self))
def transform(pattern: BranchPattern) -> Either:
"""Expand pattern into an (almost) equivalent one, but with single Either.
Example: ((-a | -b) (-c | -d)) => (-a -c | -a -d | -b -c | -b -d)
Quirks: [-a] => (-a), (-a...) => (-a -a)
"""
result = []
groups = [[pattern]]
while groups:
children = groups.pop(0)
parents = [Required, NotRequired, OptionsShortcut, Either, OneOrMore]
if any(t in map(type, children) for t in parents):
child = [c for c in children if type(c) in parents][0]
children.remove(child)
if type(child) is Either:
for c in child.children:
groups.append([c] + children)
elif type(child) is OneOrMore:
groups.append(child.children * 2 + children)
else:
groups.append(child.children + children)
else:
result.append(children)
return Either(*[Required(*e) for e in result])
TSingleMatch = Tuple[Union[int, None], Union["LeafPattern", None]]
class LeafPattern(Pattern):
"""Leaf/terminal node of a pattern tree."""
def __repr__(self) -> str:
return "%s(%r, %r)" % (self.__class__.__name__, self.name, self.value)
def single_match(self, left: list[LeafPattern]) -> TSingleMatch:
raise NotImplementedError # pragma: no cover
def flat(self, *types) -> list[LeafPattern]:
return [self] if not types or type(self) in types else []
def match(
self, left: list[LeafPattern], collected: list[Pattern] | None = None
) -> tuple[bool, list[LeafPattern], list[Pattern]]:
collected = [] if collected is None else collected
increment: Any | None = None
pos, match = self.single_match(left)
if match is None or pos is None:
return False, left, collected
left_ = left[:pos] + left[(pos + 1) :]
same_name = [a for a in collected if a.name == self.name]
if type(self.value) == int and len(same_name) > 0:
if isinstance(same_name[0].value, int):
same_name[0].value += 1
return True, left_, collected
if type(self.value) == int and not same_name:
match.value = 1
return True, left_, collected + [match]
if same_name and type(self.value) == list:
if type(match.value) == str:
increment = [match.value]
if same_name[0].value is not None and increment is not None:
if isinstance(same_name[0].value, type(increment)):
same_name[0].value += increment
return True, left_, collected
elif not same_name and type(self.value) == list:
if isinstance(match.value, str):
match.value = [match.value]
return True, left_, collected + [match]
return True, left_, collected + [match]
class BranchPattern(Pattern):
"""Branch/inner node of a pattern tree."""
def __init__(self, *children) -> None:
self.children = list(children)
def match(self, left: list[Pattern], collected: list[Pattern] | None = None) -> Any:
raise NotImplementedError # pragma: no cover
def fix(self) -> "BranchPattern":
self.fix_identities()
self.fix_repeating_arguments()
return self
def fix_identities(self, uniq: Any | None = None) -> None:
"""Make pattern-tree tips point to same object if they are equal."""
flattened = self.flat()
uniq = list(set(flattened)) if uniq is None else uniq
for i, child in enumerate(self.children):
if not hasattr(child, "children"):
assert child in uniq
self.children[i] = uniq[uniq.index(child)]
else:
child.fix_identities(uniq)
return None
def fix_repeating_arguments(self) -> BranchPattern:
"""Fix elements that should accumulate/increment values."""
either = [list(child.children) for child in transform(self).children]
for case in either:
for e in [child for child in case if case.count(child) > 1]:
if type(e) is Argument or type(e) is Option and e.argcount:
if e.value is None:
e.value = []
elif type(e.value) is not list:
e.value = cast(str, e.value)
e.value = e.value.split()
if type(e) is Command or type(e) is Option and e.argcount == 0:
e.value = 0
return self
def __repr__(self) -> str:
return "%s(%s)" % (
self.__class__.__name__,
", ".join(repr(a) for a in self.children),
)
def flat(self, *types) -> Any:
if type(self) in types:
return [self]
return sum([child.flat(*types) for child in self.children], [])
class Argument(LeafPattern):
def single_match(self, left: list[LeafPattern]) -> TSingleMatch:
for n, pattern in enumerate(left):
if type(pattern) is Argument:
return n, Argument(self.name, pattern.value)
return None, None
class Command(Argument):
def __init__(self, name: str | None, value: bool = False) -> None:
self._name, self.value = name, value
def single_match(self, left: list[LeafPattern]) -> TSingleMatch:
for n, pattern in enumerate(left):
if type(pattern) is Argument:
if pattern.value == self.name:
return n, Command(self.name, True)
else:
break
return None, None
class Option(LeafPattern):
def __init__(
self,
short: str | None = None,
longer: str | None = None,
argcount: int = 0,
value: list[str] | str | int | None = False,
) -> None:
assert argcount in (0, 1)
self.short, self.longer, self.argcount = short, longer, argcount
self.value = None if value is False and argcount else value
@classmethod
def parse(class_, option_description: str) -> Option:
short, longer, argcount, value = None, None, 0, False
options, description = re.split(
r"(?: )|$", option_description.strip(), flags=re.M, maxsplit=1
)
options = options.replace(",", " ").replace("=", " ")
for s in options.split():
if s.startswith("--"):
longer = s
elif s.startswith("-"):
short = s
else:
argcount = 1
if argcount:
matched = re.findall(r"\[default: (.*)\]", description, flags=re.I)
value = matched[0] if matched else None
return class_(short, longer, argcount, value)
def single_match(self, left: list[LeafPattern]) -> TSingleMatch:
for n, pattern in enumerate(left):
if self.name == pattern.name:
return n, pattern
return None, None
@property
def name(self) -> str | None:
return self.longer or self.short
def __repr__(self) -> str:
return "Option(%r, %r, %r, %r)" % (
self.short,
self.longer,
self.argcount,
self.value,
)
class Required(BranchPattern):
def match(self, left: list[Pattern], collected: list[Pattern] | None = None) -> Any:
collected = [] if collected is None else collected
original_collected = collected
original_left = left
for pattern in self.children:
matched, left, collected = pattern.match(left, collected)
if not matched:
return False, original_left, original_collected
return True, left, collected
class NotRequired(BranchPattern):
def match(self, left: list[Pattern], collected: list[Pattern] | None = None) -> Any:
collected = [] if collected is None else collected
for pattern in self.children:
_, left, collected = pattern.match(left, collected)
return True, left, collected
class OptionsShortcut(NotRequired):
"""Marker/placeholder for [options] shortcut."""
class OneOrMore(BranchPattern):
def match(self, left: list[Pattern], collected: list[Pattern] | None = None) -> Any:
assert len(self.children) == 1
collected = [] if collected is None else collected
original_collected = collected
original_left = left
last_left = None
matched = True
times = 0
while matched:
matched, left, collected = self.children[0].match(left, collected)
times += 1 if matched else 0
if last_left == left:
break
last_left = left
if times >= 1:
return True, left, collected
return False, original_left, original_collected
class Either(BranchPattern):
def match(self, left: list[Pattern], collected: list[Pattern] | None = None) -> Any:
collected = [] if collected is None else collected
outcomes = []
for pattern in self.children:
matched, _, _ = outcome = pattern.match(left, collected)
if matched:
outcomes.append(outcome)
if outcomes:
return min(outcomes, key=lambda outcome: len(outcome[1]))
return False, left, collected
class Tokens(list):
def __init__(
self,
source: list[str] | str,
error: Type[DocoptExit] | Type[DocoptLanguageError] = DocoptExit,
) -> None:
if isinstance(source, list):
self += source
else:
self += source.split()
self.error = error
@staticmethod
def from_pattern(source: str) -> Tokens:
source = re.sub(r"([\[\]\(\)\|]|\.\.\.)", r" \1 ", source)
fragments = [s for s in re.split(r"\s+|(\S*<.*?>)", source) if s]
return Tokens(fragments, error=DocoptLanguageError)
def move(self) -> str | None:
return self.pop(0) if len(self) else None
def current(self) -> str | None:
return self[0] if len(self) else None
def parse_longer(
tokens: Tokens, options: list[Option], argv: bool = False, more_magic: bool = False
) -> list[Pattern]:
"""longer ::= '--' chars [ ( ' ' | '=' ) chars ] ;"""
current_token = tokens.move()
if current_token is None or not current_token.startswith("--"):
raise ValueError(
f"parse_longer got what appears to be an invalid token: {current_token}"
)
longer, maybe_eq, maybe_value = current_token.partition("=")
if maybe_eq == maybe_value == "":
value = None
else:
value = maybe_value
similar = [o for o in options if o.longer and longer == o.longer]
start_collision = (
len(
[
o
for o in options
if o.longer and longer in o.longer and o.longer.startswith(longer)
]
)
> 1
)
if argv and not len(similar) and not start_collision:
similar = [
o
for o in options
if o.longer and longer in o.longer and o.longer.startswith(longer)
]
# try advanced matching
if more_magic and not similar:
corrected = [
(longer, o)
for o in options
if o.longer and levenshtein_norm(longer, o.longer) < 0.25
]
if corrected:
print(f"NB: Corrected {corrected[0][0]} to {corrected[0][1].longer}")
similar = [correct for (original, correct) in corrected]
if len(similar) > 1:
raise DocoptLanguageError(f"{longer} is not a unique prefix: {similar}?")
elif len(similar) < 1:
argcount = 1 if maybe_eq == "=" else 0
o = Option(None, longer, argcount)
options.append(o)
if tokens.error is DocoptExit:
o = Option(None, longer, argcount, value if argcount else True)
else:
o = Option(
similar[0].short, similar[0].longer, similar[0].argcount, similar[0].value
)
if o.argcount == 0:
if value is not None:
raise tokens.error("%s must not have an argument" % o.longer)
else:
if value is None:
if tokens.current() in [None, "--"]:
raise tokens.error("%s requires argument" % o.longer)
value = tokens.move()
if tokens.error is DocoptExit:
o.value = value if value is not None else True
return [o]
def parse_shorts(
tokens: Tokens, options: list[Option], more_magic: bool = False
) -> list[Pattern]:
"""shorts ::= '-' ( chars )* [ [ ' ' ] chars ] ;"""
token = tokens.move()
if token is None or not token.startswith("-") or token.startswith("--"):
raise ValueError(
f"parse_shorts got what appears to be an invalid token: {token}"
)
left = token.lstrip("-")
parsed: list[Pattern] = []
while left != "":
short, left = "-" + left[0], left[1:]
transformations: dict[str | None, Callable[[str], str]] = {None: lambda x: x}
if more_magic:
transformations["lowercase"] = lambda x: x.lower()
transformations["uppercase"] = lambda x: x.upper()
# try identity, lowercase, uppercase, iff such resolves uniquely
# (ie if upper and lowercase are not both defined)
similar: list[Option] = []
de_abbreviated = False
for transform_name, transform in transformations.items():
transformed = list(set([transform(o.short) for o in options if o.short]))
no_collisions = len(
[
o
for o in options
if o.short and transformed.count(transform(o.short)) == 1
]
) # == len(transformed)
if no_collisions:
similar = [
o
for o in options
if o.short and transform(o.short) == transform(short)
]
if similar:
if transform_name:
print(
f"NB: Corrected {short} to {similar[0].short} "
f"via {transform_name}"
)
break
# if transformations do not resolve, try abbreviations of 'longer' forms
# iff such resolves uniquely (ie if no two longer forms begin with the
# same letter)
if not similar and more_magic:
abbreviated = [
transform(o.longer[1:3])
for o in options
if o.longer and not o.short
] + [transform(o.short) for o in options if o.short and not o.longer]
nonredundantly_abbreviated_options = [
o for o in options if o.longer and abbreviated.count(short) == 1
]
no_collisions = len(nonredundantly_abbreviated_options) == len(
abbreviated
)
if no_collisions:
for o in options:
if (
not o.short
and o.longer
and transform(short) == transform(o.longer[1:3])
):
similar = [o]
print(
f"NB: Corrected {short} to {similar[0].longer} "
f"via abbreviation (case change: {transform_name})"
)
break
if len(similar):
de_abbreviated = True
break
if len(similar) > 1:
raise DocoptLanguageError(
f"{short} is specified ambiguously {len(similar)} times"
)
elif len(similar) < 1:
o = Option(short, None, 0)
options.append(o)
if tokens.error is DocoptExit:
o = Option(short, None, 0, True)
else:
if de_abbreviated:
option_short_value = None
else:
option_short_value = transform(short)
o = Option(
option_short_value,
similar[0].longer,
similar[0].argcount,
similar[0].value,
)
value = None
current_token = tokens.current()
if o.argcount != 0:
if left == "":
if current_token is None or current_token == "--":
raise tokens.error("%s requires argument" % short)
else:
value = tokens.move()
else:
value = left
left = ""
if tokens.error is DocoptExit:
o.value = value if value is not None else True
parsed.append(o)
return parsed
def parse_pattern(source: str, options: list[Option]) -> Required:
tokens = Tokens.from_pattern(source)
result = parse_expr(tokens, options)
if tokens.current() is not None:
raise tokens.error("unexpected ending: %r" % " ".join(tokens))
return Required(*result)
def parse_expr(tokens: Tokens, options: list[Option]) -> list[Pattern]:
"""expr ::= seq ( '|' seq )* ;"""
result: list[Pattern] = []
seq_0: list[Pattern] = parse_seq(tokens, options)
if tokens.current() != "|":
return seq_0
if len(seq_0) > 1:
result.append(Required(*seq_0))
else:
result += seq_0
while tokens.current() == "|":
tokens.move()
seq_1 = parse_seq(tokens, options)
if len(seq_1) > 1:
result += [Required(*seq_1)]
else:
result += seq_1
return [Either(*result)]
def parse_seq(tokens: Tokens, options: list[Option]) -> list[Pattern]:
"""seq ::= ( atom [ '...' ] )* ;"""
result: list[Pattern] = []
while tokens.current() not in [None, "]", ")", "|"]:
atom = parse_atom(tokens, options)
if tokens.current() == "...":
atom = [OneOrMore(*atom)]
tokens.move()
result += atom
return result
def parse_atom(tokens: Tokens, options: list[Option]) -> list[Pattern]:
"""atom ::= '(' expr ')' | '[' expr ']' | 'options'
| longer | shorts | argument | command ;
"""
token = tokens.current()
if not token:
return [Command(tokens.move())] # pragma: no cover
elif token in "([":
tokens.move()
matching = {"(": ")", "[": "]"}[token]
pattern = {"(": Required, "[": NotRequired}[token]
matched_pattern = pattern(*parse_expr(tokens, options))
if tokens.move() != matching:
raise tokens.error("unmatched '%s'" % token)
return [matched_pattern]
elif token == "options":
tokens.move()
return [OptionsShortcut()]
elif token.startswith("--") and token != "--":
return parse_longer(tokens, options)
elif token.startswith("-") and token not in ("-", "--"):
return parse_shorts(tokens, options)
elif token.startswith("<") and token.endswith(">") or token.isupper():
return [Argument(tokens.move())]
else:
return [Command(tokens.move())]
def parse_argv(
tokens: Tokens,
options: list[Option],
options_first: bool = False,
more_magic: bool = False,
) -> list[Pattern]:
"""Parse command-line argument vector.
If options_first:
argv ::= [ longer | shorts ]* [ argument ]* [ '--' [ argument ]* ] ;
else:
argv ::= [ longer | shorts | argument ]* [ '--' [ argument ]* ] ;
"""
def isanumber(x):
try:
float(x)
return True
except ValueError:
return False
parsed: list[Pattern] = []
current_token = tokens.current()
while current_token is not None:
if current_token == "--":
return parsed + [Argument(None, v) for v in tokens]
elif current_token.startswith("--"):
parsed += parse_longer(tokens, options, argv=True, more_magic=more_magic)
elif (
current_token.startswith("-")
and current_token != "-"
and not isanumber(current_token)
):
parsed += parse_shorts(tokens, options, more_magic=more_magic)
elif options_first:
return parsed + [Argument(None, v) for v in tokens]
else:
parsed.append(Argument(None, tokens.move()))
current_token = tokens.current()
return parsed
class DocSections(NamedTuple):
before_usage: str
usage_header: str
usage_body: str
after_usage: str
def parse_docstring_sections(docstring: str) -> DocSections:
"""Partition the docstring into the main sections.
The docstring is returned, split into a tuple of 4 pieces: text before the
usage section, the usage section header, the usage section body and text
following the usage section.
"""
usage_pattern = r"""
# Any number of lines (that don't include usage:) precede the usage section
\A(?P<before_usage>(?:(?!.*\busage:).*\n)*)
# The `usage:` section header.
^(?P<usage_header>.*\busage:)
(?P<usage_body>
# The first line of the body may follow the header without a line break:
(?:.*(?:\n|\Z))
# Any number of additional indented lines
(?:[ \t].*(?:\n|\Z))*
)
# Everything else
(?P<after_usage>(?:.|\n)*)\Z
"""
match = re.match(usage_pattern, docstring, flags=re.M | re.I | re.VERBOSE)
if not match:
raise DocoptLanguageError(
'Failed to parse doc: "usage:" section (case-insensitive) not found. '
"Check http://docopt.org/ for examples of how your doc should look."
)
before, header, body, after = match.groups()
return DocSections(before, header, body, after)
def parse_options(docstring: str) -> list[Option]:
"""Parse the option descriptions from the help text.
`docstring` is the sub-section of the overall docstring that option
descriptions should be parsed from. It must not contain the "usage:"
section, as wrapped lines in the usage pattern can be misinterpreted as
option descriptions.
Option descriptions appear below the usage patterns, They define synonymous
long and short options, options that have arguments, and the default values
of options' arguments. They look like this:
```
-v, --verbose Be more verbose
-n COUNT, --number COUNT The number of times to
do the thing [default: 42]
```
"""
option_start = r"""
# Option descriptions begin on a new line
^
# They may occur on the same line as an options: section heading
(?:.*options:)?
# They can be indented with whitespace
[ \t]*
# The description itself starts with the short or long flag (-x or --xxx)
(-\S)
"""
parts = re.split(option_start, docstring, flags=re.M | re.I | re.VERBOSE)[1:]
return [
Option.parse(start + rest) for (start, rest) in zip(parts[0::2], parts[1::2])
]
def lint_docstring(sections: DocSections):
"""Report apparent mistakes in the docstring format."""
if re.search("options:", sections.usage_body, flags=re.I):
raise DocoptLanguageError(
'Failed to parse docstring: "options:" (case-insensitive) was '
'found in "usage:" section. Use a blank line after the usage, or '
"start the next section without leading whitespace."
)
if re.search("usage:", sections.usage_body + sections.after_usage, flags=re.I):
raise DocoptLanguageError(
'Failed to parse docstring: More than one "usage:" '
"(case-insensitive) section found."
)
if sections.usage_body.strip() == "":
raise DocoptLanguageError(
'Failed to parse docstring: "usage:" section is empty.'
"Check http://docopt.org/ for examples of how your doc should look."
)
def formal_usage(usage: str) -> str:
program_name, *tokens = usage.split()
return "( " + " ".join(") | (" if s == program_name else s for s in tokens) + " )"
def extras(
default_help: bool, version: None, options: list[Pattern], docstring: str
) -> None:
if default_help and any(
(o.name in ("-h", "--help")) and o.value
for o in options
if isinstance(o, Option)
):
print(docstring.strip("\n"))
sys.exit()
if version and any(
o.name == "--version" and o.value for o in options if isinstance(o, Option)
):
print(version)
sys.exit()
class ParsedOptions(dict):
def __repr__(self):
return "{%s}" % ",\n ".join("%r: %r" % i for i in sorted(self.items()))
def __getattr__(self, name: str) -> str | bool | None:
return self.get(name) or {
name: self.get(k)
for k in self.keys()
if name in [k.lstrip("-").replace("-", "_"), k.lstrip("<").rstrip(">")]
}.get(name)
def docopt(
docstring: str,
argv: list[str] | str | None = None,
default_help: bool = True,
version: Any = None,
options_first: bool = False,
) -> ParsedOptions:
"""Parse `argv` based on command-line interface described in `docstring`.
`docopt` creates your command-line interface based on its
description that you pass as `docstring`. Such description can contain
--options, <positional-argument>, commands, which could be
[optional], (required), (mutually | exclusive) or repeated...
Parameters
----------
docstring : str
Description of your command-line interface.
argv : list of str or str, optional
Argument vector to be parsed. sys.argv[1:] is used if not
provided. If str is passed, the string is split on whitespace.
default_help : bool (default: True)
Set to False to disable automatic help on -h or --help
options.
version : any object
If passed, the object will be printed if --version is in
`argv`.
options_first : bool (default: False)
Set to True to require options precede positional arguments,
i.e. to forbid options and positional arguments intermix.
Returns
-------
arguments: dict-like
A dictionary, where keys are names of command-line elements
such as e.g. "--verbose" and "<path>", and values are the
parsed values of those elements. Also supports dot access.
Example
-------
>>> from docopt import docopt
>>> doc = '''
... Usage:
... my_program tcp <host> <port> [--timeout=<seconds>]
... my_program serial <port> [--baud=<n>] [--timeout=<seconds>]
... my_program (-h | --help | --version)
...
... Options:
... -h, --help Show this screen and exit.
... --baud=<n> Baudrate [default: 9600]
... '''
>>> argv = ['tcp', '127.0.0.1', '80', '--timeout', '30']
>>> docopt(doc, argv)
{'--baud': '9600',
'--help': False,
'--timeout': '30',
'--version': False,
'<host>': '127.0.0.1',
'<port>': '80',
'serial': False,
'tcp': True}
"""
argv = sys.argv[1:] if argv is None else argv
sections = parse_docstring_sections(docstring)
lint_docstring(sections)
DocoptExit.usage = sections.usage_header + sections.usage_body
options = [
*parse_options(sections.before_usage),
*parse_options(sections.after_usage),
]
pattern = parse_pattern(formal_usage(sections.usage_body), options)
pattern_options = set(pattern.flat(Option))
for options_shortcut in pattern.flat(OptionsShortcut):
options_shortcut.children = [
opt for opt in options if opt not in pattern_options
]
parsed_arg_vector = parse_argv(Tokens(argv), list(options), options_first)
extras(default_help, version, parsed_arg_vector, docstring)
matched, left, collected = pattern.fix().match(parsed_arg_vector)
if matched and left == []:
return ParsedOptions((a.name, a.value) for a in (pattern.flat() + collected))
if left:
raise DocoptExit(f"Warning: found unmatched (duplicate?) arguments {left}")
raise DocoptExit(collected=collected, left=left)
|