1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820
|
  
# NAME
IPC::Run - system() and background procs w/ piping, redirs, ptys (Unix, Win32)
# SYNOPSIS
## First,a command to run:
my @cat = qw( cat );
## Using run() instead of system():
use IPC::Run qw( run timeout );
run \@cat, \$in, \$out, \$err, timeout( 10 ) or die "cat: $?";
# Can do I/O to sub refs and filenames, too:
run \@cat, '<', "in.txt", \&out, \&err or die "cat: $?";
run \@cat, '<', "in.txt", '>>', "out.txt", '2>>', "err.txt";
# Redirecting using pseudo-terminals instead of pipes.
run \@cat, '<pty<', \$in, '>pty>', \$out_and_err;
## Scripting subprocesses (like Expect):
use IPC::Run qw( start pump finish timeout );
# Incrementally read from / write to scalars.
# $in is drained as it is fed to cat's stdin,
# $out accumulates cat's stdout
# $err accumulates cat's stderr
# $h is for "harness".
my $h = start \@cat, \$in, \$out, \$err, timeout( 10 );
$in .= "some input\n";
pump $h until $out =~ /input\n/g;
$in .= "some more input\n";
pump $h until $out =~ /\G.*more input\n/;
$in .= "some final input\n";
finish $h or die "cat returned $?";
warn $err if $err;
print $out; ## All of cat's output
# Piping between children
run \@cat, '|', \@gzip;
# Multiple children simultaneously (run() blocks until all
# children exit, use start() for background execution):
run \@foo1, '&', \@foo2;
# Calling \&set_up_child in the child before it executes the
# command (only works on systems with true fork() & exec())
# exceptions thrown in set_up_child() will be propagated back
# to the parent and thrown from run().
run \@cat, \$in, \$out,
init => \&set_up_child;
# Read from / write to file handles you open and close
open IN, '<in.txt' or die $!;
open OUT, '>out.txt' or die $!;
print OUT "preamble\n";
run \@cat, \*IN, \*OUT or die "cat returned $?";
print OUT "postamble\n";
close IN;
close OUT;
# Create pipes for you to read / write (like IPC::Open2 & 3).
$h = start
\@cat,
'<pipe', \*IN, # may also be a lexical filehandle e.g. \my $infh
'>pipe', \*OUT,
'2>pipe', \*ERR
or die "cat returned $?";
print IN "some input\n";
close IN;
print <OUT>, <ERR>;
finish $h;
# Mixing input and output modes
run \@cat, 'in.txt', \&catch_some_out, \*ERR_LOG;
# Other redirection constructs
run \@cat, '>&', \$out_and_err;
run \@cat, '2>&1';
run \@cat, '0<&3';
run \@cat, '<&-';
run \@cat, '3<', \$in3;
run \@cat, '4>', \$out4;
# etc.
# Passing options:
run \@cat, 'in.txt', debug => 1;
# Call this system's shell, returns TRUE on 0 exit code
# THIS IS THE OPPOSITE SENSE OF system()'s RETURN VALUE
run "cat a b c" or die "cat returned $?";
# Launch a sub process directly, no shell. Can't do redirection
# with this form, it's here to behave like system() with an
# inverted result.
$r = run "cat a b c";
# Read from a file in to a scalar
run io( "filename", 'r', \$recv );
run io( \*HANDLE, 'r', \$recv );
# DESCRIPTION
IPC::Run allows you to run and interact with child processes using files, pipes,
and pseudo-ttys. Both system()-style and scripted usages are supported and
may be mixed. Likewise, functional and OO API styles are both supported and
may be mixed.
Various redirection operators reminiscent of those seen on common Unix and DOS
command lines are provided.
Before digging in to the details a few LIMITATIONS are important enough
to be mentioned right up front:
- Win32 Support
Win32 support is working but **EXPERIMENTAL**, but does pass all relevant tests
on NT 4.0. See ["Win32 LIMITATIONS"](#win32-limitations).
- pty Support
If you need pty support, IPC::Run should work well enough most of the
time, but IO::Pty is being improved, and IPC::Run will be improved to
use IO::Pty's new features when it is released.
The basic problem is that the pty needs to initialize itself before the
parent writes to the master pty, or the data written gets lost. So
IPC::Run does a sleep(1) in the parent after forking to (hopefully) give
the child a chance to run. This is a kludge that works well on non
heavily loaded systems :(.
ptys are not supported yet under Win32, but will be emulated...
- Debugging Tip
You may use the environment variable `IPCRUNDEBUG` to see what's going on
under the hood:
$ IPCRUNDEBUG=basic myscript # prints minimal debugging
$ IPCRUNDEBUG=data myscript # prints all data reads/writes
$ IPCRUNDEBUG=details myscript # prints lots of low-level details
$ IPCRUNDEBUG=gory myscript # (Win32 only) prints data moving through
# the helper processes.
We now return you to your regularly scheduled documentation.
## Harnesses
Child processes and I/O handles are gathered in to a harness, then
started and run until the processing is finished or aborted.
## run() vs. start(); pump(); finish();
There are two modes you can run harnesses in: run() functions as an
enhanced system(), and start()/pump()/finish() allow for background
processes and scripted interactions with them.
When using run(), all data to be sent to the harness is set up in
advance (though one can feed subprocesses input from subroutine refs to
get around this limitation). The harness is run and all output is
collected from it, then any child processes are waited for:
run \@cmd, \<<IN, \$out;
blah
IN
## To precompile harnesses and run them later:
my $h = harness \@cmd, \<<IN, \$out;
blah
IN
run $h;
The background and scripting API is provided by start(), pump(), and
finish(): start() creates a harness if need be (by calling harness())
and launches any subprocesses, pump() allows you to poll them for
activity, and finish() then monitors the harnessed activities until they
complete.
## Build the harness, open all pipes, and launch the subprocesses
my $h = start \@cat, \$in, \$out;
$in = "first input\n";
## Now do I/O. start() does no I/O.
pump $h while length $in; ## Wait for all input to go
## Now do some more I/O.
$in = "second input\n";
pump $h until $out =~ /second input/;
## Clean up
finish $h or die "cat returned $?";
You can optionally compile the harness with harness() prior to
start()ing or run()ing, and you may omit start() between harness() and
pump(). You might want to do these things if you compile your harnesses
ahead of time.
## Using regexps to match output
As shown in most of the scripting examples, the read-to-scalar facility
for gathering subcommand's output is often used with regular expressions
to detect stopping points. This is because subcommand output often
arrives in dribbles and drabs, often only a character or line at a time.
This output is input for the main program and piles up in variables like
the `$out` and `$err` in our examples.
Regular expressions can be used to wait for appropriate output in
several ways. The `cat` example in the previous section demonstrates
how to pump() until some string appears in the output. Here's an
example that uses `smb` to fetch files from a remote server:
$h = harness \@smbclient, \$in, \$out;
$in = "cd /src\n";
$h->pump until $out =~ /^smb.*> \Z/m;
die "error cding to /src:\n$out" if $out =~ "ERR";
$out = '';
$in = "mget *\n";
$h->pump until $out =~ /^smb.*> \Z/m;
die "error retrieving files:\n$out" if $out =~ "ERR";
$in = "quit\n";
$h->finish;
Notice that we carefully clear $out after the first command/response
cycle? That's because IPC::Run does not delete $out when we continue,
and we don't want to trip over the old output in the second
command/response cycle.
Say you want to accumulate all the output in $out and analyze it
afterwards. Perl offers incremental regular expression matching using
the `m//gc` and pattern matching idiom and the `\G` assertion.
IPC::Run is careful not to disturb the current `pos()` value for
scalars it appends data to, so we could modify the above so as not to
destroy $out by adding a couple of `/gc` modifiers. The `/g` keeps us
from tripping over the previous prompt and the `/c` keeps us from
resetting the prior match position if the expected prompt doesn't
materialize immediately:
$h = harness \@smbclient, \$in, \$out;
$in = "cd /src\n";
$h->pump until $out =~ /^smb.*> \Z/mgc;
die "error cding to /src:\n$out" if $out =~ "ERR";
$in = "mget *\n";
$h->pump until $out =~ /^smb.*> \Z/mgc;
die "error retrieving files:\n$out" if $out =~ "ERR";
$in = "quit\n";
$h->finish;
analyze( $out );
When using this technique, you may want to preallocate $out to have
plenty of memory or you may find that the act of growing $out each time
new input arrives causes an `O(length($out)^2)` slowdown as $out grows.
Say we expect no more than 10,000 characters of input at the most. To
preallocate memory to $out, do something like:
my $out = "x" x 10_000;
$out = "";
`perl` will allocate at least 10,000 characters' worth of space, then
mark the $out as having 0 length without freeing all that yummy RAM.
## Timeouts and Timers
More than likely, you don't want your subprocesses to run forever, and
sometimes it's nice to know that they're going a little slowly.
Timeouts throw exceptions after a some time has elapsed, timers merely
cause pump() to return after some time has elapsed. Neither is
reset/restarted automatically.
Timeout objects are created by calling timeout( $interval ) and passing
the result to run(), start() or harness(). The timeout period starts
ticking just after all the child processes have been fork()ed or
spawn()ed, and are polled for expiration in run(), pump() and finish().
If/when they expire, an exception is thrown. This is typically useful
to keep a subprocess from taking too long.
If a timeout occurs in run(), all child processes will be terminated and
all file/pipe/ptty descriptors opened by run() will be closed. File
descriptors opened by the parent process and passed in to run() are not
closed in this event.
If a timeout occurs in pump(), pump\_nb(), or finish(), it's up to you to
decide whether to kill\_kill() all the children or to implement some more
graceful fallback. No I/O will be closed in pump(), pump\_nb() or
finish() by such an exception (though I/O is often closed down in those
routines during the natural course of events).
Often an exception is too harsh. timer( $interval ) creates timer
objects that merely prevent pump() from blocking forever. This can be
useful for detecting stalled I/O or printing a soothing message or "."
to pacify an anxious user.
Timeouts and timers can both be restarted at any time using the timer's
start() method (this is not the start() that launches subprocesses). To
restart a timer, you need to keep a reference to the timer:
## Start with a nice long timeout to let smbclient connect. If
## pump or finish take too long, an exception will be thrown.
my $h;
eval {
$h = harness \@smbclient, \$in, \$out, \$err, ( my $t = timeout 30 );
sleep 11; # No effect: timer not running yet
start $h;
$in = "cd /src\n";
pump $h until ! length $in;
$in = "ls\n";
## Now use a short timeout, since this should be faster
$t->start( 5 );
pump $h until ! length $in;
$t->start( 10 ); ## Give smbclient a little while to shut down.
$h->finish;
};
if ( $@ ) {
my $x = $@; ## Preserve $@ in case another exception occurs
$h->kill_kill; ## kill it gently, then brutally if need be, or just
## brutally on Win32.
die $x;
}
Timeouts and timers are _not_ checked once the subprocesses are shut
down; they will not expire in the interval between the last valid
process and when IPC::Run scoops up the processes' result codes, for
instance.
## Spawning synchronization, child exception propagation
start() pauses the parent until the child executes the command or CODE
reference and propagates any exceptions thrown (including exec()
failure) back to the parent. This has several pleasant effects: any
exceptions thrown in the child, including exec() failure, come flying
out of start() or run() as though they had occurred in the parent.
This includes exceptions your code thrown from init subs. In this
example:
eval {
run \@cmd, init => sub { die "blast it! foiled again!" };
};
print $@;
the exception "blast it! foiled again" will be thrown from the child
process (preventing the exec()) and printed by the parent.
In situations like
run \@cmd1, "|", \@cmd2, "|", \@cmd3;
@cmd1 will be initted and exec()ed before @cmd2, and @cmd2 before @cmd3.
This can save time and prevent oddball errors emitted by later commands
when earlier commands fail to execute. Note that IPC::Run doesn't start
any commands unless it can find the executables referenced by all
commands. These executables must pass both the `-f` and `-x` tests
described in [perlfunc](https://metacpan.org/pod/perlfunc).
Another nice effect is that init() subs can take their time doing things
and there will be no problems caused by a parent continuing to execute
before a child's init() routine is complete. Say the init() routine
needs to open a socket or a temp file that the parent wants to connect
to; without this synchronization, the parent will need to implement a
retry loop to wait for the child to run, since often, the parent gets a
lot of things done before the child's first timeslice is allocated.
This is also quite necessary for pseudo-tty initialization, which needs
to take place before the parent writes to the child via pty. Writes
that occur before the pty is set up can get lost.
A final, minor, nicety is that debugging output from the child will be
emitted before the parent continues on, making for much clearer debugging
output in complex situations.
The only drawback I can conceive of is that the parent can't continue to
operate while the child is being initted. If this ever becomes a
problem in the field, we can implement an option to avoid this behavior,
but I don't expect it to.
**Win32**: executing CODE references isn't supported on Win32, see
["Win32 LIMITATIONS"](#win32-limitations) for details.
## Syntax
run(), start(), and harness() can all take a harness specification
as input. A harness specification is either a single string to be passed
to the systems' shell:
run "echo 'hi there'";
or a list of commands, io operations, and/or timers/timeouts to execute.
Consecutive commands must be separated by a pipe operator '|' or an '&'.
External commands are passed in as array references or [IPC::Run::Win32Process](https://metacpan.org/pod/IPC%3A%3ARun%3A%3AWin32Process)
objects. On systems supporting fork(), Perl code may be passed in as subs:
run \@cmd;
run \@cmd1, '|', \@cmd2;
run \@cmd1, '&', \@cmd2;
run \&sub1;
run \&sub1, '|', \&sub2;
run \&sub1, '&', \&sub2;
'|' pipes the stdout of \\@cmd1 the stdin of \\@cmd2, just like a
shell pipe. '&' does not. Child processes to the right of a '&'
will have their stdin closed unless it's redirected-to.
[IPC::Run::IO](https://metacpan.org/pod/IPC%3A%3ARun%3A%3AIO) objects may be passed in as well, whether or not
child processes are also specified:
run io( "infile", ">", \$in ), io( "outfile", "<", \$in );
as can [IPC::Run::Timer](https://metacpan.org/pod/IPC%3A%3ARun%3A%3ATimer) objects:
run \@cmd, io( "outfile", "<", \$in ), timeout( 10 );
Commands may be followed by scalar, sub, or i/o handle references for
redirecting
child process input & output:
run \@cmd, \undef, \$out;
run \@cmd, \$in, \$out;
run \@cmd1, \&in, '|', \@cmd2, \*OUT;
run \@cmd1, \*IN, '|', \@cmd2, \&out;
This is known as succinct redirection syntax, since run(), start()
and harness(), figure out which file descriptor to redirect and how.
File descriptor 0 is presumed to be an input for
the child process, all others are outputs. The assumed file
descriptor always starts at 0, unless the command is being piped to,
in which case it starts at 1.
To be explicit about your redirects, or if you need to do more complex
things, there's also a redirection operator syntax:
run \@cmd, '<', \undef, '>', \$out;
run \@cmd, '<', \undef, '>&', \$out_and_err;
run(
\@cmd1,
'<', \$in,
'|', \@cmd2,
\$out
);
Operator syntax is required if you need to do something other than simple
redirection to/from scalars or subs, like duping or closing file descriptors
or redirecting to/from a named file. The operators are covered in detail
below.
After each \\@cmd (or \\&foo), parsing begins in succinct mode and toggles to
operator syntax mode when an operator (ie plain scalar, not a ref) is seen.
Once in
operator syntax mode, parsing only reverts to succinct mode when a '|' or
'&' is seen.
In succinct mode, each parameter after the \\@cmd specifies what to
do with the next highest file descriptor. These File descriptor start
with 0 (stdin) unless stdin is being piped to (`'|', \@cmd`), in which
case they start with 1 (stdout). Currently, being on the left of
a pipe (`\@cmd, \$out, \$err, '|'`) does _not_ cause stdout to be
skipped, though this may change since it's not as DWIMerly as it
could be. Only stdin is assumed to be an
input in succinct mode, all others are assumed to be outputs.
If no piping or redirection is specified for a child, it will inherit
the parent's open file handles as dictated by your system's
close-on-exec behavior and the $^F flag, except that processes after a
'&' will not inherit the parent's stdin. Also note that $^F does not
affect file descriptors obtained via POSIX, since it only applies to
full-fledged Perl file handles. Such processes will have their stdin
closed unless it has been redirected-to.
If you want to close a child processes stdin, you may do any of:
run \@cmd, \undef;
run \@cmd, \"";
run \@cmd, '<&-';
run \@cmd, '0<&-';
Redirection is done by placing redirection specifications immediately
after a command or child subroutine:
run \@cmd1, \$in, '|', \@cmd2, \$out;
run \@cmd1, '<', \$in, '|', \@cmd2, '>', \$out;
If you omit the redirection operators, descriptors are counted
starting at 0. Descriptor 0 is assumed to be input, all others
are outputs. A leading '|' consumes descriptor 0, so this
works as expected.
run \@cmd1, \$in, '|', \@cmd2, \$out;
The parameter following a redirection operator can be a scalar ref,
a subroutine ref, a file name, an open filehandle, or a closed
filehandle.
If it's a scalar ref, the child reads input from or sends output to
that variable:
$in = "Hello World.\n";
run \@cat, \$in, \$out;
print $out;
Scalars used in incremental (start()/pump()/finish()) applications are treated
as queues: input is removed from input scalars, resulting in them dwindling
to '', and output is appended to output scalars. This is not true of
harnesses run() in batch mode.
It's usually wise to append new input to be sent to the child to the input
queue, and you'll often want to zap output queues to '' before pumping.
$h = start \@cat, \$in;
$in = "line 1\n";
pump $h;
$in .= "line 2\n";
pump $h;
$in .= "line 3\n";
finish $h;
The final call to finish() must be there: it allows the child process(es)
to run to completion and waits for their exit values.
# OBSTINATE CHILDREN
Interactive applications are usually optimized for human use. This
can help or hinder trying to interact with them through modules like
IPC::Run. Frequently, programs alter their behavior when they detect
that stdin, stdout, or stderr are not connected to a tty, assuming that
they are being run in batch mode. Whether this helps or hurts depends
on which optimizations change. And there's often no way of telling
what a program does in these areas other than trial and error and
occasionally, reading the source. This includes different versions
and implementations of the same program.
All hope is not lost, however. Most programs behave in reasonably
tractable manners, once you figure out what it's trying to do.
Here are some of the issues you might need to be aware of.
- fflush()ing stdout and stderr
This lets the user see stdout and stderr immediately. Many programs
undo this optimization if stdout is not a tty, making them harder to
manage by things like IPC::Run.
Many programs decline to fflush stdout or stderr if they do not
detect a tty there. Some ftp commands do this, for instance.
If this happens to you, look for a way to force interactive behavior,
like a command line switch or command. If you can't, you will
need to use a pseudo terminal ('<pty<' and '>pty>').
- false prompts
Interactive programs generally do not guarantee that output from user
commands won't contain a prompt string. For example, your shell prompt
might be a '$', and a file named '$' might be the only file in a directory
listing.
This can make it hard to guarantee that your output parser won't be fooled
into early termination of results.
To help work around this, you can see if the program can alter it's
prompt, and use something you feel is never going to occur in actual
practice.
You should also look for your prompt to be the only thing on a line:
pump $h until $out =~ /^<SILLYPROMPT>\s?\z/m;
(use `(?!\n)\Z` in place of `\z` on older perls).
You can also take the approach that IPC::ChildSafe takes and emit a
command with known output after each 'real' command you issue, then
look for this known output. See new\_appender() and new\_chunker() for
filters that can help with this task.
If it's not convenient or possibly to alter a prompt or use a known
command/response pair, you might need to autodetect the prompt in case
the local version of the child program is different then the one
you tested with, or if the user has control over the look & feel of
the prompt.
- Refusing to accept input unless stdin is a tty.
Some programs, for security reasons, will only accept certain types
of input from a tty. su, notable, will not prompt for a password unless
it's connected to a tty.
If this is your situation, use a pseudo terminal ('<pty<' and '>pty>').
- Not prompting unless connected to a tty.
Some programs don't prompt unless stdin or stdout is a tty. See if you can
turn prompting back on. If not, see if you can come up with a command that
you can issue after every real command and look for it's output, as
IPC::ChildSafe does. There are two filters included with IPC::Run that
can help with doing this: appender and chunker (see new\_appender() and
new\_chunker()).
- Different output format when not connected to a tty.
Some commands alter their formats to ease machine parsability when they
aren't connected to a pipe. This is actually good, but can be surprising.
# PSEUDO TERMINALS
On systems providing pseudo terminals under /dev, IPC::Run can use IO::Pty
(available on CPAN) to provide a terminal environment to subprocesses.
This is necessary when the subprocess really wants to think it's connected
to a real terminal.
## CAVEATS
Pseudo-terminals are not pipes, though they are similar. Here are some
differences to watch out for.
- Echoing
Sending to stdin will cause an echo on stdout, which occurs before each
line is passed to the child program. There is currently no way to
disable this, although the child process can and should disable it for
things like passwords.
- Shutdown
IPC::Run cannot close a pty until all output has been collected. This
means that it is not possible to send an EOF to stdin by half-closing
the pty, as we can when using a pipe to stdin.
This means that you need to send the child process an exit command or
signal, or run() / finish() will time out. Be careful not to expect a
prompt after sending the exit command.
- Command line editing
Some subprocesses, notable shells that depend on the user's prompt
settings, will reissue the prompt plus the command line input so far
once for each character.
- '>pty>' means '&>pty>', not '1>pty>'
The pseudo terminal redirects both stdout and stderr unless you specify
a file descriptor. If you want to grab stderr separately, do this:
start \@cmd, '<pty<', \$in, '>pty>', \$out, '2>', \$err;
- stdin, stdout, and stderr not inherited
Child processes harnessed to a pseudo terminal have their stdin, stdout,
and stderr completely closed before any redirection operators take
effect. This casts off the bonds of the controlling terminal. This is
not done when using pipes.
Right now, this affects all children in a harness that has a pty in use,
even if that pty would not affect a particular child. That's a bug and
will be fixed. Until it is, it's best not to mix-and-match children.
## Redirection Operators
Operator SHNP Description
======== ==== ===========
<, N< SHN Redirects input to a child's fd N (0 assumed)
>, N> SHN Redirects output from a child's fd N (1 assumed)
>>, N>> SHN Like '>', but appends to scalars or named files
>&, &> SHN Redirects stdout & stderr from a child process
<pty, N<pty S Like '<', but uses a pseudo-tty instead of a pipe
>pty, N>pty S Like '>', but uses a pseudo-tty instead of a pipe
N<&M Dups input fd N to input fd M
M>&N Dups output fd N to input fd M
N<&- Closes fd N
<pipe, N<pipe P Pipe opens H for caller to read, write, close.
>pipe, N>pipe P Pipe opens H for caller to read, write, close.
'N' and 'M' are placeholders for integer file descriptor numbers. The
terms 'input' and 'output' are from the child process's perspective.
The SHNP field indicates what parameters an operator can take:
S: \$scalar or \&function references. Filters may be used with
these operators (and only these).
H: \*HANDLE or IO::Handle for caller to open, and close
N: "file name".
P: \*HANDLE or lexical filehandle opened by IPC::Run as the parent end of a pipe, but read
and written to and closed by the caller (like IPC::Open3).
- Redirecting input: \[n\]<, \[n\]<pipe
You can input the child reads on file descriptor number n to come from a
scalar variable, subroutine, file handle, or a named file. If stdin
is not redirected, the parent's stdin is inherited.
run \@cat, \undef ## Closes child's stdin immediately
or die "cat returned $?";
run \@cat, \$in;
run \@cat, \<<TOHERE;
blah
TOHERE
run \@cat, \&input; ## Calls &input, feeding data returned
## to child's. Closes child's stdin
## when undef is returned.
Redirecting from named files requires you to use the input
redirection operator:
run \@cat, '<.profile';
run \@cat, '<', '.profile';
open IN, "<foo";
run \@cat, \*IN;
run \@cat, *IN{IO};
The form used second example here is the safest,
since filenames like "0" and "&more\\n" won't confuse &run:
You can't do either of
run \@a, *IN; ## INVALID
run \@a, '<', *IN; ## BUGGY: Reads file named like "*main::A"
because perl passes a scalar containing a string that
looks like "\*main::A" to &run, and &run can't tell the difference
between that and a redirection operator or a file name. &run guarantees
that any scalar you pass after a redirection operator is a file name.
If your child process will take input from file descriptors other
than 0 (stdin), you can use a redirection operator with any of the
valid input forms (scalar ref, sub ref, etc.):
run \@cat, '3<', \$in3;
When redirecting input from a scalar ref, the scalar ref is
used as a queue. This allows you to use &harness and pump() to
feed incremental bits of input to a coprocess. See ["Coprocesses"](#coprocesses)
below for more information.
The <pipe operator opens the write half of a pipe on the filehandle
glob reference it takes as an argument:
$h = start \@cat, '<pipe', \*IN;
print IN "hello world\n";
pump $h;
close IN;
finish $h;
Unlike the other '<' operators, IPC::Run does nothing further with
it: you are responsible for it. The previous example is functionally
equivalent to:
pipe( \*R, \*IN ) or die $!;
$h = start \@cat, '<', \*IN;
print IN "hello world\n";
pump $h;
close IN;
finish $h;
This is like the behavior of IPC::Open2 and IPC::Open3.
**Win32**: The handle returned is actually a socket handle, so you can
use select() on it.
- Redirecting output: \[n\]>, \[n\]>>, \[n\]>&\[m\], \[n\]>pipe
You can redirect any output the child emits
to a scalar variable, subroutine, file handle, or file name. You
can have &run truncate or append to named files or scalars. If
you are redirecting stdin as well, or if the command is on the
receiving end of a pipeline ('|'), you can omit the redirection
operator:
@ls = ( 'ls' );
run \@ls, \undef, \$out
or die "ls returned $?";
run \@ls, \undef, \&out; ## Calls &out each time some output
## is received from the child's
## when undef is returned.
run \@ls, \undef, '2>ls.err';
run \@ls, '2>', 'ls.err';
The two parameter form guarantees that the filename
will not be interpreted as a redirection operator:
run \@ls, '>', "&more";
run \@ls, '2>', ">foo\n";
You can pass file handles you've opened for writing:
open( *OUT, ">out.txt" );
open( *ERR, ">err.txt" );
run \@cat, \*OUT, \*ERR;
Passing a scalar reference and a code reference requires a little
more work, but allows you to capture all of the output in a scalar
or each piece of output by a callback:
These two do the same things:
run( [ 'ls' ], '2>', sub { $err_out .= $_[0] } );
does the same basic thing as:
run( [ 'ls' ], '2>', \$err_out );
The subroutine will be called each time some data is read from the child.
The >pipe operator is different in concept than the other '>' operators,
although it's syntax is similar:
$h = start \@cat, $in, '>pipe', \*OUT, '2>pipe', \*ERR;
$in = "hello world\n";
finish $h;
print <OUT>;
print <ERR>;
close OUT;
close ERR;
causes two pipe to be created, with one end attached to cat's stdout
and stderr, respectively, and the other left open on OUT and ERR, so
that the script can manually
read(), select(), etc. on them. This is like
the behavior of IPC::Open2 and IPC::Open3.
**Win32**: The handle returned is actually a socket handle, so you can
use select() on it.
- Duplicating output descriptors: >&m, n>&m
This duplicates output descriptor number n (default is 1 if n is omitted)
from descriptor number m.
- Duplicating input descriptors: <&m, n<&m
This duplicates input descriptor number n (default is 0 if n is omitted)
from descriptor number m
- Closing descriptors: <&-, 3<&-
This closes descriptor number n (default is 0 if n is omitted). The
following commands are equivalent:
run \@cmd, \undef;
run \@cmd, '<&-';
run \@cmd, '<in.txt', '<&-';
Doing
run \@cmd, \$in, '<&-'; ## SIGPIPE recipe.
is dangerous: the parent will get a SIGPIPE if $in is not empty.
- Redirecting both stdout and stderr: &>, >&, &>pipe, >pipe&
The following pairs of commands are equivalent:
run \@cmd, '>&', \$out; run \@cmd, '>', \$out, '2>&1';
run \@cmd, '>&', 'out.txt'; run \@cmd, '>', 'out.txt', '2>&1';
etc.
File descriptor numbers are not permitted to the left or the right of
these operators, and the '&' may occur on either end of the operator.
The '&>pipe' and '>pipe&' variants behave like the '>pipe' operator, except
that both stdout and stderr write to the created pipe.
- Redirection Filters
Both input redirections and output redirections that use scalars or
subs as endpoints may have an arbitrary number of filter subs placed
between them and the child process. This is useful if you want to
receive output in chunks, or if you want to massage each chunk of
data sent to the child. To use this feature, you must use operator
syntax:
run(
\@cmd
'<', \&in_filter_2, \&in_filter_1, $in,
'>', \&out_filter_1, \&in_filter_2, $out,
);
This capability is not provided for IO handles or named files.
Two filters are provided by IPC::Run: appender and chunker. Because
these may take an argument, you need to use the constructor functions
new\_appender() and new\_chunker() rather than using \\& syntax:
run(
\@cmd
'<', new_appender( "\n" ), $in,
'>', new_chunker, $out,
);
## Just doing I/O
If you just want to do I/O to a handle or file you open yourself, you
may specify a filehandle or filename instead of a command in the harness
specification:
run io( "filename", '>', \$recv );
$h = start io( $io, '>', \$recv );
$h = harness \@cmd, '&', io( "file", '<', \$send );
## Options
Options are passed in as name/value pairs:
run \@cat, \$in, debug => 1;
If you pass the debug option, you may want to pass it in first, so you
can see what parsing is going on:
run debug => 1, \@cat, \$in;
- debug
Enables debugging output in parent and child. Debugging info is emitted
to the STDERR that was present when IPC::Run was first `use()`ed (it's
`dup()`ed out of the way so that it can be redirected in children without
having debugging output emitted on it).
# RETURN VALUES
harness() and start() return a reference to an IPC::Run harness. This is
blessed in to the IPC::Run package, so you may make later calls to
functions as members if you like:
$h = harness( ... );
$h->start;
$h->pump;
$h->finish;
$h = start( .... );
$h->pump;
...
Of course, using method call syntax lets you deal with any IPC::Run
subclasses that might crop up, but don't hold your breath waiting for
any.
run() and finish() return TRUE when all subcommands exit with a 0 result
code. **This is the opposite of perl's system() command**.
All routines raise exceptions (via die()) when error conditions are
recognized. A non-zero command result is not treated as an error
condition, since some commands are tests whose results are reported
in their exit codes.
# ROUTINES
> - run
>
> Run takes a harness or harness specification and runs it, pumping
> all input to the child(ren), closing the input pipes when no more
> input is available, collecting all output that arrives, until the
> pipes delivering output are closed, then waiting for the children to
> exit and reaping their result codes.
>
> You may think of `run( ... )` as being like
>
> start( ... )->finish();
>
> , though there is one subtle difference: run() does not
> set \\$input\_scalars to '' like finish() does. If an exception is thrown
> from run(), all children will be killed off "gently", and then "annihilated"
> if they do not go gently (in to that dark night. sorry).
>
> If any exceptions are thrown, this does a ["kill\_kill"](#kill_kill) before propagating
> them.
>
> - signal
>
> ## To send it a specific signal by name ("USR1"):
> signal $h, "USR1";
> $h->signal ( "USR1" );
>
> If $signal is provided and defined, sends a signal to all child processes. Try
> not to send numeric signals, use `"KILL"` instead of `9`, for instance.
> Numeric signals aren't portable.
>
> Throws an exception if $signal is undef.
>
> This will _not_ clean up the harness, `finish` it if you kill it.
>
> Normally TERM kills a process gracefully (this is what the command line utility
> `kill` does by default), INT is sent by one of the keys `^C`, `Backspace` or
> `<Del>`, and `QUIT` is used to kill a process and make it coredump.
>
> The `HUP` signal is often used to get a process to "restart", rereading
> config files, and `USR1` and `USR2` for really application-specific things.
>
> Often, running `kill -l` (that's a lower case "L") on the command line will
> list the signals present on your operating system.
>
> **WARNING**: The signal subsystem is not at all portable. We \*may\* offer
> to simulate `TERM` and `KILL` on some operating systems, submit code
> to me if you want this.
>
> **WARNING 2**: Up to and including perl v5.6.1, doing almost anything in a
> signal handler could be dangerous. The most safe code avoids all
> mallocs and system calls, usually by preallocating a flag before
> entering the signal handler, altering the flag's value in the
> handler, and responding to the changed value in the main system:
>
> my $got_usr1 = 0;
> sub usr1_handler { ++$got_signal }
>
> $SIG{USR1} = \&usr1_handler;
> while () { sleep 1; print "GOT IT" while $got_usr1--; }
>
> Even this approach is perilous if ++ and -- aren't atomic on your system
> (I've never heard of this on any modern CPU large enough to run perl).
>
> - kill\_kill
>
> ## To kill off a process:
> $h->kill_kill;
> kill_kill $h;
>
> ## To specify the grace period other than 30 seconds:
> kill_kill $h, grace => 5;
>
> ## To send QUIT instead of KILL if a process refuses to die:
> kill_kill $h, coup_d_grace => "QUIT";
>
> Sends a `TERM`, waits for all children to exit for up to 30 seconds, then
> sends a `KILL` to any that survived the `TERM`.
>
> Will wait for up to 30 more seconds for the OS to successfully `KILL` the
> processes.
>
> The 30 seconds may be overridden by setting the `grace` option, this
> overrides both timers.
>
> The harness is then cleaned up.
>
> The doubled name indicates that this function may kill again and avoids
> colliding with the core Perl `kill` function.
>
> Returns a 1 if the `TERM` was sufficient, or a 0 if `KILL` was
> required. Throws an exception if `KILL` did not permit the children
> to be reaped.
>
> **NOTE**: The grace period is actually up to 1 second longer than that
> given. This is because the granularity of `time` is 1 second. Let me
> know if you need finer granularity, we can leverage Time::HiRes here.
>
> **Win32**: Win32 does not know how to send real signals, so `TERM` is
> a full-force kill on Win32. Thus all talk of grace periods, etc. do
> not apply to Win32.
>
> - harness
>
> Takes a harness specification and returns a harness. This harness is
> blessed in to IPC::Run, allowing you to use method call syntax for
> run(), start(), et al if you like.
>
> harness() is provided so that you can pre-build harnesses if you
> would like to, but it's not required..
>
> You may proceed to run(), start() or pump() after calling harness() (pump()
> calls start() if need be). Alternatively, you may pass your
> harness specification to run() or start() and let them harness() for
> you. You can't pass harness specifications to pump(), though.
>
> - close\_terminal
>
> This is used as (or in) an init sub to cast off the bonds of a controlling
> terminal. It must precede all other redirection ops that affect
> STDIN, STDOUT, or STDERR to be guaranteed effective.
>
> - start
>
> $h = start(
> \@cmd, \$in, \$out, ...,
> timeout( 30, name => "process timeout" ),
> $stall_timeout = timeout( 10, name => "stall timeout" ),
> );
>
> $h = start \@cmd, '<', \$in, '|', \@cmd2, ...;
>
> start() accepts a harness or harness specification and returns a harness
> after building all of the pipes and launching (via fork()/exec(), or, maybe
> someday, spawn()) all the child processes. It does not send or receive any
> data on the pipes, see pump() and finish() for that.
>
> You may call harness() and then pass it's result to start() if you like,
> but you only need to if it helps you structure or tune your application.
> If you do call harness(), you may skip start() and proceed directly to
> pump.
>
> start() also starts all timers in the harness. See [IPC::Run::Timer](https://metacpan.org/pod/IPC%3A%3ARun%3A%3ATimer)
> for more information.
>
> start() flushes STDOUT and STDERR to help you avoid duplicate output.
> It has no way of asking Perl to flush all your open filehandles, so
> you are going to need to flush any others you have open. Sorry.
>
> Here's how if you don't want to alter the state of $| for your
> filehandle:
>
> $ofh = select HANDLE; $of = $|; $| = 1; $| = $of; select $ofh;
>
> If you don't mind leaving output unbuffered on HANDLE, you can do
> the slightly shorter
>
> $ofh = select HANDLE; $| = 1; select $ofh;
>
> Or, you can use IO::Handle's flush() method:
>
> use IO::Handle;
> flush HANDLE;
>
> Perl needs the equivalent of C's fflush( (FILE \*)NULL ).
>
> - adopt
>
> Experimental feature. NOT FUNCTIONAL YET, NEED TO CLOSE FDS BETTER IN CHILDREN. SEE t/adopt.t for a test suite.
>
> - pump
>
> pump $h;
> $h->pump;
>
> Pump accepts a single parameter harness. It blocks until it delivers some
> input or receives some output. It returns TRUE if there is still input or
> output to be done, FALSE otherwise.
>
> pump() will automatically call start() if need be, so you may call harness()
> then proceed to pump() if that helps you structure your application.
>
> If pump() is called after all harnessed activities have completed, a "process
> ended prematurely" exception to be thrown. This allows for simple scripting
> of external applications without having to add lots of error handling code at
> each step of the script:
>
> $h = harness \@smbclient, \$in, \$out, $err;
>
> $in = "cd /foo\n";
> $h->pump until $out =~ /^smb.*> \Z/m;
> die "error cding to /foo:\n$out" if $out =~ "ERR";
> $out = '';
>
> $in = "mget *\n";
> $h->pump until $out =~ /^smb.*> \Z/m;
> die "error retrieving files:\n$out" if $out =~ "ERR";
>
> $h->finish;
>
> warn $err if $err;
>
> - pump\_nb
>
> pump_nb $h;
> $h->pump_nb;
>
> "pump() non-blocking", pumps if anything's ready to be pumped, returns
> immediately otherwise. This is useful if you're doing some long-running
> task in the foreground, but don't want to starve any child processes.
>
> - pumpable
>
> Returns TRUE if calling pump() won't throw an immediate "process ended
> prematurely" exception. This means that there are open I/O channels or
> active processes. May yield the parent processes' time slice for 0.01
> second if all pipes are to the child and all are paused. In this case
> we can't tell if the child is dead, so we yield the processor and
> then attempt to reap the child in a nonblocking way.
>
> - reap\_nb
>
> Attempts to reap child processes, but does not block.
>
> Does not currently take any parameters, one day it will allow specific
> children to be reaped.
>
> Only call this from a signal handler if your `perl` is recent enough
> to have safe signal handling (5.6.1 did not, IIRC, but it was being discussed
> on perl5-porters). Calling this (or doing any significant work) in a signal
> handler on older `perl`s is asking for seg faults.
>
> - finish
>
> This must be called after the last start() or pump() call for a harness,
> or your system will accumulate defunct processes and you may "leak"
> file descriptors.
>
> finish() returns TRUE if all children returned 0 (and were not signaled and did
> not coredump, ie ! $?), and FALSE otherwise (this is like run(), and the
> opposite of system()).
>
> Once a harness has been finished, it may be run() or start()ed again,
> including by pump()s auto-start.
>
> If this throws an exception rather than a normal exit, the harness may
> be left in an unstable state, it's best to kill the harness to get rid
> of all the child processes, etc.
>
> Specifically, if a timeout expires in finish(), finish() will not
> kill all the children. Call `$h->kill_kill` in this case if you care.
> This differs from the behavior of ["run"](#run).
>
> - result
>
> $h->result;
>
> Returns the first non-zero result code (ie $? >> 8). See ["full\_result"](#full_result) to
> get the $? value for a child process.
>
> To get the result of a particular child, do:
>
> $h->result( 0 ); # first child's $? >> 8
> $h->result( 1 ); # second child
>
> or
>
> ($h->results)[0]
> ($h->results)[1]
>
> Returns undef if no child processes were spawned and no child number was
> specified. Throws an exception if an out-of-range child number is passed.
>
> - results
>
> Returns a list of child exit values. See ["full\_results"](#full_results) if you want to
> know if a signal killed the child.
>
> Throws an exception if the harness is not in a finished state.
>
> - full\_result
>
> $h->full_result;
>
> Returns the first non-zero $?. See ["result"](#result) to get the first $? >> 8
> value for a child process.
>
> To get the result of a particular child, do:
>
> $h->full_result( 0 ); # first child's $?
> $h->full_result( 1 ); # second child
>
> or
>
> ($h->full_results)[0]
> ($h->full_results)[1]
>
> Returns undef if no child processes were spawned and no child number was
> specified. Throws an exception if an out-of-range child number is passed.
>
> - full\_results
>
> Returns a list of child exit values as returned by `wait`. See ["results"](#results)
> if you don't care about coredumps or signals.
>
> Throws an exception if the harness is not in a finished state.
# FILTERS
These filters are used to modify input our output between a child
process and a scalar or subroutine endpoint.
- binary
run \@cmd, ">", binary, \$out;
run \@cmd, ">", binary, \$out; ## Any TRUE value to enable
run \@cmd, ">", binary 0, \$out; ## Any FALSE value to disable
This is a constructor for a "binmode" "filter" that tells IPC::Run to keep
the carriage returns that would ordinarily be edited out for you (binmode
is usually off). This is not a real filter, but an option masquerading as
a filter.
It's not named "binmode" because you're likely to want to call Perl's binmode
in programs that are piping binary data around.
- new\_chunker
This breaks a stream of data in to chunks, based on an optional
scalar or regular expression parameter. The default is the Perl
input record separator in $/, which is a newline be default.
run \@cmd, '>', new_chunker, \&lines_handler;
run \@cmd, '>', new_chunker( "\r\n" ), \&lines_handler;
Because this uses $/ by default, you should always pass in a parameter
if you are worried about other code (modules, etc) modifying $/.
If this filter is last in a filter chain that dumps in to a scalar,
the scalar must be set to '' before a new chunk will be written to it.
As an example of how a filter like this can be written, here's a
chunker that splits on newlines:
sub line_splitter {
my ( $in_ref, $out_ref ) = @_;
return 0 if length $$out_ref;
return input_avail && do {
while (1) {
if ( $$in_ref =~ s/\A(.*?\n)// ) {
$$out_ref .= $1;
return 1;
}
my $hmm = get_more_input;
unless ( defined $hmm ) {
$$out_ref = $$in_ref;
$$in_ref = '';
return length $$out_ref ? 1 : 0;
}
return 0 if $hmm eq 0;
}
}
};
- new\_appender
This appends a fixed string to each chunk of data read from the source
scalar or sub. This might be useful if you're writing commands to a
child process that always must end in a fixed string, like "\\n":
run( \@cmd,
'<', new_appender( "\n" ), \&commands,
);
Here's a typical filter sub that might be created by new\_appender():
sub newline_appender {
my ( $in_ref, $out_ref ) = @_;
return input_avail && do {
$$out_ref = join( '', $$out_ref, $$in_ref, "\n" );
$$in_ref = '';
1;
}
};
- new\_string\_source
TODO: Needs confirmation. Was previously undocumented. in this module.
This is a filter which is exportable. Returns a sub which appends the data passed in to the output buffer and returns 1 if data was appended. 0 if it was an empty string and undef if no data was passed.
NOTE: Any additional variables passed to new\_string\_source will be passed to the sub every time it's called and appended to the output.
- new\_string\_sink
TODO: Needs confirmation. Was previously undocumented.
This is a filter which is exportable. Returns a sub which pops the data out of the input stream and pushes it onto the string.
- io
Takes a filename or filehandle, a redirection operator, optional filters,
and a source or destination (depends on the redirection operator). Returns
an IPC::Run::IO object suitable for harness()ing (including via start()
or run()).
This is shorthand for
require IPC::Run::IO;
... IPC::Run::IO->new(...) ...
- timer
$h = start( \@cmd, \$in, \$out, $t = timer( 5 ) );
pump $h until $out =~ /expected stuff/ || $t->is_expired;
Instantiates a non-fatal timer. pump() returns once each time a timer
expires. Has no direct effect on run(), but you can pass a subroutine
to fire when the timer expires.
See ["timeout"](#timeout) for building timers that throw exceptions on
expiration.
See ["timer" in IPC::Run::Timer](https://metacpan.org/pod/IPC%3A%3ARun%3A%3ATimer#timer) for details.
- timeout
$h = start( \@cmd, \$in, \$out, $t = timeout( 5 ) );
pump $h until $out =~ /expected stuff/;
Instantiates a timer that throws an exception when it expires.
If you don't provide an exception, a default exception that matches
/^IPC::Run: .\*timed out/ is thrown by default. You can pass in your own
exception scalar or reference:
$h = start(
\@cmd, \$in, \$out,
$t = timeout( 5, exception => 'slowpoke' ),
);
or set the name used in debugging message and in the default exception
string:
$h = start(
\@cmd, \$in, \$out,
timeout( 50, name => 'process timer' ),
$stall_timer = timeout( 5, name => 'stall timer' ),
);
pump $h until $out =~ /started/;
$in = 'command 1';
$stall_timer->start;
pump $h until $out =~ /command 1 finished/;
$in = 'command 2';
$stall_timer->start;
pump $h until $out =~ /command 2 finished/;
$in = 'very slow command 3';
$stall_timer->start( 10 );
pump $h until $out =~ /command 3 finished/;
$stall_timer->start( 5 );
$in = 'command 4';
pump $h until $out =~ /command 4 finished/;
$stall_timer->reset; # Prevent restarting or expirng
finish $h;
See ["timer"](#timer) for building non-fatal timers.
See ["timer" in IPC::Run::Timer](https://metacpan.org/pod/IPC%3A%3ARun%3A%3ATimer#timer) for details.
# FILTER IMPLEMENTATION FUNCTIONS
These functions are for use from within filters.
- input\_avail
Returns TRUE if input is available. If none is available, then
&get\_more\_input is called and its result is returned.
This is usually used in preference to &get\_more\_input so that the
calling filter removes all data from the $in\_ref before more data
gets read in to $in\_ref.
`input_avail` is usually used as part of a return expression:
return input_avail && do {
## process the input just gotten
1;
};
This technique allows input\_avail to return the undef or 0 that a
filter normally returns when there's no input to process. If a filter
stores intermediate values, however, it will need to react to an
undef:
my $got = input_avail;
if ( ! defined $got ) {
## No more input ever, flush internal buffers to $out_ref
}
return $got unless $got;
## Got some input, move as much as need be
return 1 if $added_to_out_ref;
- get\_more\_input
This is used to fetch more input in to the input variable. It returns
undef if there will never be any more input, 0 if there is none now,
but there might be in the future, and TRUE if more input was gotten.
`get_more_input` is usually used as part of a return expression,
see ["input\_avail"](#input_avail) for more information.
# TODO
These will be addressed as needed and as time allows.
Stall timeout.
Expose a list of child process objects. When I do this,
each child process is likely to be blessed into IPC::Run::Proc.
$kid->abort(), $kid->kill(), $kid->signal( $num\_or\_name ).
Write tests for /(full\_)?results?/ subs.
Currently, pump() and run() only work on systems where select() works on the
filehandles returned by pipe(). This does \*not\* include ActiveState on Win32,
although it does work on cygwin under Win32 (thought the tests whine a bit).
I'd like to rectify that, suggestions and patches welcome.
Likewise start() only fully works on fork()/exec() machines (well, just
fork() if you only ever pass perl subs as subprocesses). There's
some scaffolding for calling Open3::spawn\_with\_handles(), but that's
untested, and not that useful with limited select().
Support for `\@sub_cmd` as an argument to a command which
gets replaced with /dev/fd or the name of a temporary file containing foo's
output. This is like <(sub\_cmd ...) found in bash and csh (IIRC).
Allow multiple harnesses to be combined as independent sets of processes
in to one 'meta-harness'.
Allow a harness to be passed in place of an \\@cmd. This would allow
multiple harnesses to be aggregated.
Ability to add external file descriptors w/ filter chains and endpoints.
Ability to add timeouts and timing generators (i.e. repeating timeouts).
High resolution timeouts.
# Win32 LIMITATIONS
- argument-passing rules are program-specific
Win32 programs receive all arguments in a single "command line" string.
IPC::Run assembles this string so programs using [standard command line parsing
rules](https://docs.microsoft.com/en-us/cpp/cpp/main-function-command-line-args#parsing-c-command-line-arguments)
will see an `argv` that matches the array reference specifying the command.
Some programs use different rules to parse their command line. Notable examples
include `cmd.exe`, `cscript.exe`, and Cygwin programs called from non-Cygwin
programs. Use [IPC::Run::Win32Process](https://metacpan.org/pod/IPC%3A%3ARun%3A%3AWin32Process) to call these and other nonstandard
programs.
- batch files
Properly escaping a batch file argument depends on how the script will use that
argument, because some uses experience multiple levels of caret (escape
character) removal. Avoid calling batch files with arguments, particularly when
the argument values originate outside your program or contain non-alphanumeric
characters. Perl scripts and PowerShell scripts are sound alternatives. If you
do use batch file arguments, IPC::Run escapes them so the batch file can pass
them, unquoted, to a program having standard command line parsing rules. If the
batch file enables delayed environment variable expansion, it must disable that
feature before expanding its arguments. For example, if `foo.cmd` contains
`perl %*`, `run ['foo.cmd', @list]` will create a Perl process in which
`@ARGV` matches `@list`. Prepending a `setlocal enabledelayedexpansion` line
would make the batch file malfunction, silently. Another silent-malfunction
example is `run ['outer.bat', @list]` for `outer.bat` containing `foo.cmd
%*`.
- Fails on Win9X
If you want Win9X support, you'll have to debug it or fund me because I
don't use that system any more. The Win32 subsysem has been extended to
use temporary files in simple run() invocations and these may actually
work on Win9X too, but I don't have time to work on it.
- May deadlock on Win2K (but not WinNT4 or WinXPPro)
Spawning more than one subprocess on Win2K causes a deadlock I haven't
figured out yet, but simple uses of run() often work. Passes all tests
on WinXPPro and WinNT.
- no support yet for <pty< and >pty>
These are likely to be implemented as "<" and ">" with binmode on, not
sure.
- no support for file descriptors higher than 2 (stderr)
Win32 only allows passing explicit fds 0, 1, and 2. If you really, really need to pass file handles, us Win32API:: GetOsFHandle() or ::FdGetOsFHandle() to
get the integer handle and pass it to the child process using the command
line, environment, stdin, intermediary file, or other IPC mechanism. Then
use that handle in the child (Win32API.pm provides ways to reconstitute
Perl file handles from Win32 file handles).
- no support for subroutine subprocesses (CODE refs)
Can't fork(), so the subroutines would have no context, and closures certainly
have no meaning
Perhaps with Win32 fork() emulation, this can be supported in a limited
fashion, but there are other very serious problems with that: all parent
fds get dup()ed in to the thread emulating the forked process, and that
keeps the parent from being able to close all of the appropriate fds.
- no support for init => sub {} routines.
Win32 processes are created from scratch, there is no way to do an init
routine that will affect the running child. Some limited support might
be implemented one day, do chdir() and %ENV changes can be made.
- signals
Win32 does not fully support signals. signal() is likely to cause errors
unless sending a signal that Perl emulates, and `kill_kill()` is immediately
fatal (there is no grace period).
- `$?` cannot represent all Win32 exit codes
Prefer `full_result( ... )`, `result( ... )`, or other IPC::Run methods.
- helper processes
IPC::Run uses helper processes, one per redirected file, to adapt between the
anonymous pipe connected to the child and the TCP socket connected to the
parent. This is a waste of resources and will change in the future to either
use threads (instead of helper processes) or a WaitForMultipleObjects call
(instead of select). Please contact me if you can help with the
WaitForMultipleObjects() approach; I haven't figured out how to get at it
without C code.
- shutdown pause
There seems to be a pause of up to 1 second between when a child program exits
and the corresponding sockets indicate that they are closed in the parent.
Not sure why.
- binmode
binmode is not supported yet. The underpinnings are implemented, just ask
if you need it.
- IPC::Run::IO
IPC::Run::IO objects can be used on Unix to read or write arbitrary files. On
Win32, they will need to use the same helper processes to adapt from
non-select()able filehandles to select()able ones (or perhaps
WaitForMultipleObjects() will work with them, not sure).
- startup race conditions
There seems to be an occasional race condition between child process startup
and pipe closings. It seems like if the child is not fully created by the time
CreateProcess returns and we close the TCP socket being handed to it, the
parent socket can also get closed. This is seen with the Win32 pumper
applications, not the "real" child process being spawned.
I assume this is because the kernel hasn't gotten around to incrementing the
reference count on the child's end (since the child was slow in starting), so
the parent's closing of the child end causes the socket to be closed, thus
closing the parent socket.
Being a race condition, it's hard to reproduce, but I encountered it while
testing this code on a drive share to a samba box. In this case, it takes
t/run.t a long time to spawn it's child processes (the parent hangs in the
first select for several seconds until the child emits any debugging output).
I have not seen it on local drives, and can't reproduce it at will,
unfortunately. The symptom is a "bad file descriptor in select()" error, and,
by turning on debugging, it's possible to see that select() is being called on
a no longer open file descriptor that was returned from the \_socket() routine
in Win32Helper. There's a new confess() that checks for this ("PARENT\_HANDLE
no longer open"), but I haven't been able to reproduce it (typically).
# LIMITATIONS
On Unix, requires a system that supports `waitpid( $pid, WNOHANG )` so
it can tell if a child process is still running.
PTYs don't seem to be non-blocking on some versions of Solaris. Here's a
test script contributed by Borislav Deianov <borislav@ensim.com> to see
if you have the problem. If it dies, you have the problem.
#!/usr/bin/perl
use IPC::Run qw(run);
use Fcntl;
use IO::Pty;
sub makecmd {
return ['perl', '-e',
'<STDIN>, print "\n" x '.$_[0].'; while(<STDIN>){last if /end/}'];
}
#pipe R, W;
#fcntl(W, F_SETFL, O_NONBLOCK);
#while (syswrite(W, "\n", 1)) { $pipebuf++ };
#print "pipe buffer size is $pipebuf\n";
my $pipebuf=4096;
my $in = "\n" x ($pipebuf * 2) . "end\n";
my $out;
$SIG{ALRM} = sub { die "Never completed!\n" };
print "reading from scalar via pipe...";
alarm( 2 );
run(makecmd($pipebuf * 2), '<', \$in, '>', \$out);
alarm( 0 );
print "done\n";
print "reading from code via pipe... ";
alarm( 2 );
run(makecmd($pipebuf * 3), '<', sub { $t = $in; undef $in; $t}, '>', \$out);
alarm( 0 );
print "done\n";
$pty = IO::Pty->new();
$pty->blocking(0);
$slave = $pty->slave();
while ($pty->syswrite("\n", 1)) { $ptybuf++ };
print "pty buffer size is $ptybuf\n";
$in = "\n" x ($ptybuf * 3) . "end\n";
print "reading via pty... ";
alarm( 2 );
run(makecmd($ptybuf * 3), '<pty<', \$in, '>', \$out);
alarm(0);
print "done\n";
No support for ';', '&&', '||', '{ ... }', etc: use perl's, since run()
returns TRUE when the command exits with a 0 result code.
Does not provide shell-like string interpolation.
No support for `cd`, `setenv`, or `export`: do these in an init() sub
run(
\cmd,
...
init => sub {
chdir $dir or die $!;
$ENV{FOO}='BAR'
}
);
Timeout calculation does not allow absolute times, or specification of
days, months, etc.
**WARNING:** Function coprocesses (`run \&foo, ...`) suffer from two
limitations. The first is that it is difficult to close all filehandles the
child inherits from the parent, since there is no way to scan all open
FILEHANDLEs in Perl and it both painful and a bit dangerous to close all open
file descriptors with `POSIX::close()`. Painful because we can't tell which
fds are open at the POSIX level, either, so we'd have to scan all possible fds
and close any that we don't want open (normally `exec()` closes any
non-inheritable but we don't `exec()` for &sub processes.
The second problem is that Perl's DESTROY subs and other on-exit cleanup gets
run in the child process. If objects are instantiated in the parent before the
child is forked, the DESTROY will get run once in the parent and once in
the child. When coprocess subs exit, POSIX::\_exit is called to work around this,
but it means that objects that are still referred to at that time are not
cleaned up. So setting package vars or closure vars to point to objects that
rely on DESTROY to affect things outside the process (files, etc), will
lead to bugs.
I goofed on the syntax: "<pipe" vs. "<pty<" and ">filename" are both
oddities.
# TODO
- Allow one harness to "adopt" another:
$new_h = harness \@cmd2;
$h->adopt( $new_h );
- Close all filehandles not explicitly marked to stay open.
The problem with this one is that there's no good way to scan all open
FILEHANDLEs in Perl, yet you don't want child processes inheriting handles
willy-nilly.
# INSPIRATION
Well, select() and waitpid() badly needed wrapping, and open3() isn't
open-minded enough for me.
The shell-like API inspired by a message Russ Allbery sent to perl5-porters,
which included:
I've thought for some time that it would be
nice to have a module that could handle full Bourne shell pipe syntax
internally, with fork and exec, without ever invoking a shell. Something
that you could give things like:
pipeopen (PIPE, [ qw/cat file/ ], '|', [ 'analyze', @args ], '>&3');
Message ylln51p2b6.fsf@windlord.stanford.edu, on 2000/02/04.
# SUPPORT
Bugs should always be submitted via the GitHub bug tracker
[https://github.com/toddr/IPC-Run/issues](https://github.com/toddr/IPC-Run/issues)
# AUTHORS
Adam Kennedy <adamk@cpan.org>
Barrie Slaymaker <barries@slaysys.com>
# COPYRIGHT
Some parts copyright 2008 - 2009 Adam Kennedy.
Copyright 1999 Barrie Slaymaker.
You may distribute under the terms of either the GNU General Public
License or the Artistic License, as specified in the README file.
|