1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775
|
/*
Copyright (c) 1990-2008 Info-ZIP. All rights reserved.
See the accompanying file LICENSE, version 2007-Mar-04 or later
(the contents of which are also included in unzip.h) for terms of use.
If, for some reason, all these files are missing, the Info-ZIP license
also may be found at: ftp://ftp.info-zip.org/pub/infozip/license.html
*/
/* inflate.c -- by Mark Adler
version c17e, 30 Mar 2007 */
/* Copyright history:
- Starting with UnZip 5.41 of 16-April-2000, this source file
is covered by the Info-Zip LICENSE cited above.
- Prior versions of this source file, found in UnZip source packages
up to UnZip 5.40, were put in the public domain.
The original copyright note by Mark Adler was:
"You can do whatever you like with this source file,
though I would prefer that if you modify it and
redistribute it that you include comments to that effect
with your name and the date. Thank you."
History:
vers date who what
---- --------- -------------- ------------------------------------
a ~~ Feb 92 M. Adler used full (large, one-step) lookup table
b1 21 Mar 92 M. Adler first version with partial lookup tables
b2 21 Mar 92 M. Adler fixed bug in fixed-code blocks
b3 22 Mar 92 M. Adler sped up match copies, cleaned up some
b4 25 Mar 92 M. Adler added prototypes; removed window[] (now
is the responsibility of unzip.h--also
changed name to slide[]), so needs diffs
for unzip.c and unzip.h (this allows
compiling in the small model on MSDOS);
fixed cast of q in huft_build();
b5 26 Mar 92 M. Adler got rid of unintended macro recursion.
b6 27 Mar 92 M. Adler got rid of nextbyte() routine. fixed
bug in inflate_fixed().
c1 30 Mar 92 M. Adler removed lbits, dbits environment variables.
changed BMAX to 16 for explode. Removed
OUTB usage, and replaced it with flush()--
this was a 20% speed improvement! Added
an explode.c (to replace unimplod.c) that
uses the huft routines here. Removed
register union.
c2 4 Apr 92 M. Adler fixed bug for file sizes a multiple of 32k.
c3 10 Apr 92 M. Adler reduced memory of code tables made by
huft_build significantly (factor of two to
three).
c4 15 Apr 92 M. Adler added NOMEMCPY do kill use of memcpy().
worked around a Turbo C optimization bug.
c5 21 Apr 92 M. Adler added the WSIZE #define to allow reducing
the 32K window size for specialized
applications.
c6 31 May 92 M. Adler added some typecasts to eliminate warnings
c7 27 Jun 92 G. Roelofs added some more typecasts (444: MSC bug).
c8 5 Oct 92 J-l. Gailly added ifdef'd code to deal with PKZIP bug.
c9 9 Oct 92 M. Adler removed a memory error message (~line 416).
c10 17 Oct 92 G. Roelofs changed ULONG/UWORD/byte to ulg/ush/uch,
removed old inflate, renamed inflate_entry
to inflate, added Mark's fix to a comment.
c10.5 14 Dec 92 M. Adler fix up error messages for incomplete trees.
c11 2 Jan 93 M. Adler fixed bug in detection of incomplete
tables, and removed assumption that EOB is
the longest code (bad assumption).
c12 3 Jan 93 M. Adler make tables for fixed blocks only once.
c13 5 Jan 93 M. Adler allow all zero length codes (pkzip 2.04c
outputs one zero length code for an empty
distance tree).
c14 12 Mar 93 M. Adler made inflate.c standalone with the
introduction of inflate.h.
c14b 16 Jul 93 G. Roelofs added (unsigned) typecast to w at 470.
c14c 19 Jul 93 J. Bush changed v[N_MAX], l[288], ll[28x+3x] arrays
to static for Amiga.
c14d 13 Aug 93 J-l. Gailly de-complicatified Mark's c[*p++]++ thing.
c14e 8 Oct 93 G. Roelofs changed memset() to memzero().
c14f 22 Oct 93 G. Roelofs renamed quietflg to qflag; made Trace()
conditional; added inflate_free().
c14g 28 Oct 93 G. Roelofs changed l/(lx+1) macro to pointer (Cray bug)
c14h 7 Dec 93 C. Ghisler huft_build() optimizations.
c14i 9 Jan 94 A. Verheijen set fixed_t{d,l} to NULL after freeing;
G. Roelofs check NEXTBYTE macro for EOF.
c14j 23 Jan 94 G. Roelofs removed Ghisler "optimizations"; ifdef'd
EOF check.
c14k 27 Feb 94 G. Roelofs added some typecasts to avoid warnings.
c14l 9 Apr 94 G. Roelofs fixed split comments on preprocessor lines
to avoid bug in Encore compiler.
c14m 7 Jul 94 P. Kienitz modified to allow assembler version of
inflate_codes() (define ASM_INFLATECODES)
c14n 22 Jul 94 G. Roelofs changed fprintf to macro for DLL versions
c14o 23 Aug 94 C. Spieler added a newline to a debug statement;
G. Roelofs added another typecast to avoid MSC warning
c14p 4 Oct 94 G. Roelofs added (voidp *) cast to free() argument
c14q 30 Oct 94 G. Roelofs changed fprintf macro to MESSAGE()
c14r 1 Nov 94 G. Roelofs fixed possible redefinition of CHECK_EOF
c14s 7 May 95 S. Maxwell OS/2 DLL globals stuff incorporated;
P. Kienitz "fixed" ASM_INFLATECODES macro/prototype
c14t 18 Aug 95 G. Roelofs added UZinflate() to use zlib functions;
changed voidp to zvoid; moved huft_build()
and huft_free() to end of file
c14u 1 Oct 95 G. Roelofs moved G into definition of MESSAGE macro
c14v 8 Nov 95 P. Kienitz changed ASM_INFLATECODES to use a regular
call with __G__ instead of a macro
c15 3 Aug 96 M. Adler fixed bomb-bug on random input data (Adobe)
c15b 24 Aug 96 M. Adler more fixes for random input data
c15c 28 Mar 97 G. Roelofs changed USE_ZLIB fatal exit code from
PK_MEM2 to PK_MEM3
c16 20 Apr 97 J. Altman added memzero(v[]) in huft_build()
c16b 29 Mar 98 C. Spieler modified DLL code for slide redirection
c16c 04 Apr 99 C. Spieler fixed memory leaks when processing gets
stopped because of input data errors
c16d 05 Jul 99 C. Spieler take care of FLUSH() return values and
stop processing in case of errors
c17 31 Dec 00 C. Spieler added preliminary support for Deflate64
c17a 04 Feb 01 C. Spieler complete integration of Deflate64 support
c17b 16 Feb 02 C. Spieler changed type of "extra bits" arrays and
corresponding huft_build() parameter e from
ush into uch, to save space
c17c 9 Mar 02 C. Spieler fixed NEEDBITS() "read beyond EOF" problem
with CHECK_EOF enabled
c17d 23 Jul 05 C. Spieler fixed memory leaks in inflate_dynamic()
when processing invalid compressed literal/
distance table data
c17e 30 Mar 07 C. Spieler in inflate_dynamic(), initialize tl and td
to prevent freeing unallocated huft tables
when processing invalid compressed data and
hitting premature EOF, do not reuse td as
temp work ptr during tables decoding
*/
/*
Inflate deflated (PKZIP's method 8 compressed) data. The compression
method searches for as much of the current string of bytes (up to a
length of 258) in the previous 32K bytes. If it doesn't find any
matches (of at least length 3), it codes the next byte. Otherwise, it
codes the length of the matched string and its distance backwards from
the current position. There is a single Huffman code that codes both
single bytes (called "literals") and match lengths. A second Huffman
code codes the distance information, which follows a length code. Each
length or distance code actually represents a base value and a number
of "extra" (sometimes zero) bits to get to add to the base value. At
the end of each deflated block is a special end-of-block (EOB) literal/
length code. The decoding process is basically: get a literal/length
code; if EOB then done; if a literal, emit the decoded byte; if a
length then get the distance and emit the referred-to bytes from the
sliding window of previously emitted data.
There are (currently) three kinds of inflate blocks: stored, fixed, and
dynamic. The compressor outputs a chunk of data at a time and decides
which method to use on a chunk-by-chunk basis. A chunk might typically
be 32K to 64K, uncompressed. If the chunk is uncompressible, then the
"stored" method is used. In this case, the bytes are simply stored as
is, eight bits per byte, with none of the above coding. The bytes are
preceded by a count, since there is no longer an EOB code.
If the data are compressible, then either the fixed or dynamic methods
are used. In the dynamic method, the compressed data are preceded by
an encoding of the literal/length and distance Huffman codes that are
to be used to decode this block. The representation is itself Huffman
coded, and so is preceded by a description of that code. These code
descriptions take up a little space, and so for small blocks, there is
a predefined set of codes, called the fixed codes. The fixed method is
used if the block ends up smaller that way (usually for quite small
chunks); otherwise the dynamic method is used. In the latter case, the
codes are customized to the probabilities in the current block and so
can code it much better than the pre-determined fixed codes can.
The Huffman codes themselves are decoded using a multi-level table
lookup, in order to maximize the speed of decoding plus the speed of
building the decoding tables. See the comments below that precede the
lbits and dbits tuning parameters.
GRR: return values(?)
0 OK
1 incomplete table
2 bad input
3 not enough memory
the following return codes are passed through from FLUSH() errors
50 (PK_DISK) "overflow of output space"
80 (IZ_CTRLC) "canceled by user's request"
*/
/*
Notes beyond the 1.93a appnote.txt:
1. Distance pointers never point before the beginning of the output
stream.
2. Distance pointers can point back across blocks, up to 32k away.
3. There is an implied maximum of 7 bits for the bit length table and
15 bits for the actual data.
4. If only one code exists, then it is encoded using one bit. (Zero
would be more efficient, but perhaps a little confusing.) If two
codes exist, they are coded using one bit each (0 and 1).
5. There is no way of sending zero distance codes--a dummy must be
sent if there are none. (History: a pre 2.0 version of PKZIP would
store blocks with no distance codes, but this was discovered to be
too harsh a criterion.) Valid only for 1.93a. 2.04c does allow
zero distance codes, which is sent as one code of zero bits in
length.
6. There are up to 286 literal/length codes. Code 256 represents the
end-of-block. Note however that the static length tree defines
288 codes just to fill out the Huffman codes. Codes 286 and 287
cannot be used though, since there is no length base or extra bits
defined for them. Similarily, there are up to 30 distance codes.
However, static trees define 32 codes (all 5 bits) to fill out the
Huffman codes, but the last two had better not show up in the data.
7. Unzip can check dynamic Huffman blocks for complete code sets.
The exception is that a single code would not be complete (see #4).
8. The five bits following the block type is really the number of
literal codes sent minus 257.
9. Length codes 8,16,16 are interpreted as 13 length codes of 8 bits
(1+6+6). Therefore, to output three times the length, you output
three codes (1+1+1), whereas to output four times the same length,
you only need two codes (1+3). Hmm.
10. In the tree reconstruction algorithm, Code = Code + Increment
only if BitLength(i) is not zero. (Pretty obvious.)
11. Correction: 4 Bits: # of Bit Length codes - 4 (4 - 19)
12. Note: length code 284 can represent 227-258, but length code 285
really is 258. The last length deserves its own, short code
since it gets used a lot in very redundant files. The length
258 is special since 258 - 3 (the min match length) is 255.
13. The literal/length and distance code bit lengths are read as a
single stream of lengths. It is possible (and advantageous) for
a repeat code (16, 17, or 18) to go across the boundary between
the two sets of lengths.
14. The Deflate64 (PKZIP method 9) variant of the compression algorithm
differs from "classic" deflate in the following 3 aspect:
a) The size of the sliding history window is expanded to 64 kByte.
b) The previously unused distance codes #30 and #31 code distances
from 32769 to 49152 and 49153 to 65536. Both codes take 14 bits
of extra data to determine the exact position in their 16 kByte
range.
c) The last lit/length code #285 gets a different meaning. Instead
of coding a fixed maximum match length of 258, it is used as a
"generic" match length code, capable of coding any length from
3 (min match length + 0) to 65538 (min match length + 65535).
This means that the length code #285 takes 16 bits (!) of uncoded
extra data, added to a fixed min length of 3.
Changes a) and b) would have been transparent for valid deflated
data, but change c) requires to switch decoder configurations between
Deflate and Deflate64 modes.
*/
#define PKZIP_BUG_WORKAROUND /* PKZIP 1.93a problem--live with it */
/*
inflate.h must supply the uch slide[WSIZE] array, the zvoid typedef
(void if (void *) is accepted, else char) and the NEXTBYTE,
FLUSH() and memzero macros. If the window size is not 32K, it
should also define WSIZE. If INFMOD is defined, it can include
compiled functions to support the NEXTBYTE and/or FLUSH() macros.
There are defaults for NEXTBYTE and FLUSH() below for use as
examples of what those functions need to do. Normally, you would
also want FLUSH() to compute a crc on the data. inflate.h also
needs to provide these typedefs:
typedef unsigned char uch;
typedef unsigned short ush;
typedef unsigned long ulg;
This module uses the external functions malloc() and free() (and
probably memset() or bzero() in the memzero() macro). Their
prototypes are normally found in <string.h> and <stdlib.h>.
*/
#define __INFLATE_C /* identifies this source module */
/* #define DEBUG */
#define INFMOD /* tell inflate.h to include code to be compiled */
#include "inflate.h"
/* marker for "unused" huft code, and corresponding check macro */
#define INVALID_CODE 99
#define IS_INVALID_CODE(c) ((c) == INVALID_CODE)
#ifndef WSIZE /* default is 32K resp. 64K */
# ifdef USE_DEFLATE64
# define WSIZE 65536L /* window size--must be a power of two, and */
# else /* at least 64K for PKZip's deflate64 method */
# define WSIZE 0x8000 /* window size--must be a power of two, and */
# endif /* at least 32K for zip's deflate method */
#endif
/* some buffer counters must be capable of holding 64k for Deflate64 */
#if (defined(USE_DEFLATE64) && defined(INT_16BIT))
# define UINT_D64 ulg
#else
# define UINT_D64 unsigned
#endif
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
# define wsize G._wsize /* wsize is a variable */
#else
# define wsize WSIZE /* wsize is a constant */
#endif
#ifndef NEXTBYTE /* default is to simply get a byte from stdin */
# define NEXTBYTE getchar()
#endif
#ifndef MESSAGE /* only used twice, for fixed strings--NOT general-purpose */
# define MESSAGE(str,len,flag) fprintf(stderr,(char *)(str))
#endif
#ifndef FLUSH /* default is to simply write the buffer to stdout */
# define FLUSH(n) \
(((extent)fwrite(redirSlide, 1, (extent)(n), stdout) == (extent)(n)) ? \
0 : PKDISK)
#endif
/* Warning: the fwrite above might not work on 16-bit compilers, since
0x8000 might be interpreted as -32,768 by the library function. When
support for Deflate64 is enabled, the window size is 64K and the
simple fwrite statement is definitely broken for 16-bit compilers. */
#ifndef Trace
# ifdef DEBUG
# define Trace(x) fprintf x
# else
# define Trace(x)
# endif
#endif
/*---------------------------------------------------------------------------*/
#ifdef USE_ZLIB
/* Beginning with zlib version 1.2.0, a new inflate callback interface is
provided that allows tighter integration of the zlib inflate service
into unzip's extraction framework.
The advantages are:
- uses the windows buffer supplied by the unzip code; this saves one
copy process between zlib's internal decompression buffer and unzip's
post-decompression output buffer and improves performance.
- does not pull in unused checksum code (adler32).
The preprocessor flag NO_ZLIBCALLBCK can be set to force usage of the
old zlib 1.1.x interface, for testing purpose.
*/
#ifdef USE_ZLIB_INFLATCB
# undef USE_ZLIB_INFLATCB
#endif
#if (defined(ZLIB_VERNUM) && ZLIB_VERNUM >= 0x1200 && !defined(NO_ZLIBCALLBCK))
# define USE_ZLIB_INFLATCB 1
#else
# define USE_ZLIB_INFLATCB 0
#endif
/* Check for incompatible combinations of zlib and Deflate64 support. */
#if defined(USE_DEFLATE64)
# if !USE_ZLIB_INFLATCB
#error Deflate64 is incompatible with traditional (pre-1.2.x) zlib interface!
# else
/* The Deflate64 callback function in the framework of zlib 1.2.x requires
the inclusion of the unsupported infback9 header file:
*/
# include "infback9.h"
# endif
#endif /* USE_DEFLATE64 */
#if USE_ZLIB_INFLATCB
static unsigned zlib_inCB OF((void FAR *pG, unsigned char FAR * FAR * pInbuf));
static int zlib_outCB OF((void FAR *pG, unsigned char FAR *outbuf,
unsigned outcnt));
static unsigned zlib_inCB(pG, pInbuf)
void FAR *pG;
unsigned char FAR * FAR * pInbuf;
{
*pInbuf = G.inbuf;
return fillinbuf(__G);
}
static int zlib_outCB(pG, outbuf, outcnt)
void FAR *pG;
unsigned char FAR *outbuf;
unsigned outcnt;
{
#ifdef FUNZIP
return flush(__G__ (ulg)(outcnt));
#else
return ((G.mem_mode) ? memflush(__G__ outbuf, (ulg)(outcnt))
: flush(__G__ outbuf, (ulg)(outcnt), 0));
#endif
}
#endif /* USE_ZLIB_INFLATCB */
/*
GRR: return values for both original inflate() and UZinflate()
0 OK
1 incomplete table(?)
2 bad input
3 not enough memory
*/
/**************************/
/* Function UZinflate() */
/**************************/
int UZinflate(__G__ is_defl64)
__GDEF
int is_defl64;
/* decompress an inflated entry using the zlib routines */
{
int retval = 0; /* return code: 0 = "no error" */
int err=Z_OK;
#if USE_ZLIB_INFLATCB
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
if (G.redirect_slide)
wsize = G.redirect_size, redirSlide = G.redirect_buffer;
else
wsize = WSIZE, redirSlide = slide;
#endif
if (!G.inflInit) {
/* local buffer for efficiency */
ZCONST char *zlib_RtVersion = zlibVersion();
/* only need to test this stuff once */
if ((zlib_RtVersion[0] != ZLIB_VERSION[0]) ||
(zlib_RtVersion[2] != ZLIB_VERSION[2])) {
Info(slide, 0x21, ((char *)slide,
"error: incompatible zlib version (expected %s, found %s)\n",
ZLIB_VERSION, zlib_RtVersion));
return 3;
} else if (strcmp(zlib_RtVersion, ZLIB_VERSION) != 0)
Info(slide, 0x21, ((char *)slide,
"warning: different zlib version (expected %s, using %s)\n",
ZLIB_VERSION, zlib_RtVersion));
G.dstrm.zalloc = (alloc_func)Z_NULL;
G.dstrm.zfree = (free_func)Z_NULL;
G.inflInit = 1;
}
#ifdef USE_DEFLATE64
if (is_defl64)
{
Trace((stderr, "initializing inflate9()\n"));
err = inflateBack9Init(&G.dstrm, redirSlide);
if (err == Z_MEM_ERROR)
return 3;
else if (err != Z_OK) {
Trace((stderr, "oops! (inflateBack9Init() err = %d)\n", err));
return 2;
}
G.dstrm.next_in = G.inptr;
G.dstrm.avail_in = G.incnt;
err = inflateBack9(&G.dstrm, zlib_inCB, &G, zlib_outCB, &G);
if (err != Z_STREAM_END) {
if (err == Z_DATA_ERROR || err == Z_STREAM_ERROR) {
Trace((stderr, "oops! (inflateBack9() err = %d)\n", err));
retval = 2;
} else if (err == Z_MEM_ERROR) {
retval = 3;
} else if (err == Z_BUF_ERROR) {
Trace((stderr, "oops! (inflateBack9() err = %d)\n", err));
if (G.dstrm.next_in == Z_NULL) {
/* input failure */
Trace((stderr, " inflateBack9() input failure\n"));
retval = 2;
} else {
/* output write failure */
retval = (G.disk_full != 0 ? PK_DISK : IZ_CTRLC);
}
} else {
Trace((stderr, "oops! (inflateBack9() err = %d)\n", err));
retval = 2;
}
}
if (G.dstrm.next_in != NULL) {
G.inptr = (uch *)G.dstrm.next_in;
G.incnt = G.dstrm.avail_in;
}
err = inflateBack9End(&G.dstrm);
if (err != Z_OK) {
Trace((stderr, "oops! (inflateBack9End() err = %d)\n", err));
if (retval == 0)
retval = 2;
}
}
else
#endif /* USE_DEFLATE64 */
{
/* For the callback interface, inflate initialization has to
be called before each decompression call.
*/
{
unsigned i;
int windowBits;
/* windowBits = log2(wsize) */
for (i = (unsigned)wsize, windowBits = 0;
!(i & 1); i >>= 1, ++windowBits);
if ((unsigned)windowBits > (unsigned)15)
windowBits = 15;
else if (windowBits < 8)
windowBits = 8;
Trace((stderr, "initializing inflate()\n"));
err = inflateBackInit(&G.dstrm, windowBits, redirSlide);
if (err == Z_MEM_ERROR)
return 3;
else if (err != Z_OK) {
Trace((stderr, "oops! (inflateBackInit() err = %d)\n", err));
return 2;
}
}
G.dstrm.next_in = G.inptr;
G.dstrm.avail_in = G.incnt;
err = inflateBack(&G.dstrm, zlib_inCB, &G, zlib_outCB, &G);
if (err != Z_STREAM_END) {
if (err == Z_DATA_ERROR || err == Z_STREAM_ERROR) {
Trace((stderr, "oops! (inflateBack() err = %d)\n", err));
retval = 2;
} else if (err == Z_MEM_ERROR) {
retval = 3;
} else if (err == Z_BUF_ERROR) {
Trace((stderr, "oops! (inflateBack() err = %d)\n", err));
if (G.dstrm.next_in == Z_NULL) {
/* input failure */
Trace((stderr, " inflateBack() input failure\n"));
retval = 2;
} else {
/* output write failure */
retval = (G.disk_full != 0 ? PK_DISK : IZ_CTRLC);
}
} else {
Trace((stderr, "oops! (inflateBack() err = %d)\n", err));
retval = 2;
}
}
if (G.dstrm.next_in != NULL) {
G.inptr = (uch *)G.dstrm.next_in;
G.incnt = G.dstrm.avail_in;
}
err = inflateBackEnd(&G.dstrm);
if (err != Z_OK) {
Trace((stderr, "oops! (inflateBackEnd() err = %d)\n", err));
if (retval == 0)
retval = 2;
}
}
#else /* !USE_ZLIB_INFLATCB */
int repeated_buf_err;
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
if (G.redirect_slide)
wsize = G.redirect_size, redirSlide = G.redirect_buffer;
else
wsize = WSIZE, redirSlide = slide;
#endif
G.dstrm.next_out = redirSlide;
G.dstrm.avail_out = wsize;
G.dstrm.next_in = G.inptr;
G.dstrm.avail_in = G.incnt;
if (!G.inflInit) {
unsigned i;
int windowBits;
/* local buffer for efficiency */
ZCONST char *zlib_RtVersion = zlibVersion();
/* only need to test this stuff once */
if (zlib_RtVersion[0] != ZLIB_VERSION[0]) {
Info(slide, 0x21, ((char *)slide,
"error: incompatible zlib version (expected %s, found %s)\n",
ZLIB_VERSION, zlib_RtVersion));
return 3;
} else if (strcmp(zlib_RtVersion, ZLIB_VERSION) != 0)
Info(slide, 0x21, ((char *)slide,
"warning: different zlib version (expected %s, using %s)\n",
ZLIB_VERSION, zlib_RtVersion));
/* windowBits = log2(wsize) */
for (i = (unsigned)wsize, windowBits = 0;
!(i & 1); i >>= 1, ++windowBits);
if ((unsigned)windowBits > (unsigned)15)
windowBits = 15;
else if (windowBits < 8)
windowBits = 8;
G.dstrm.zalloc = (alloc_func)Z_NULL;
G.dstrm.zfree = (free_func)Z_NULL;
Trace((stderr, "initializing inflate()\n"));
err = inflateInit2(&G.dstrm, -windowBits);
if (err == Z_MEM_ERROR)
return 3;
else if (err != Z_OK)
Trace((stderr, "oops! (inflateInit2() err = %d)\n", err));
G.inflInit = 1;
}
#ifdef FUNZIP
while (err != Z_STREAM_END) {
#else /* !FUNZIP */
while (G.csize > 0) {
Trace((stderr, "first loop: G.csize = %ld\n", G.csize));
#endif /* ?FUNZIP */
while (G.dstrm.avail_out > 0) {
err = inflate(&G.dstrm, Z_PARTIAL_FLUSH);
if (err == Z_DATA_ERROR) {
retval = 2; goto uzinflate_cleanup_exit;
} else if (err == Z_MEM_ERROR) {
retval = 3; goto uzinflate_cleanup_exit;
} else if (err != Z_OK && err != Z_STREAM_END)
Trace((stderr, "oops! (inflate(first loop) err = %d)\n", err));
#ifdef FUNZIP
if (err == Z_STREAM_END) /* "END-of-entry-condition" ? */
#else /* !FUNZIP */
if (G.csize <= 0L) /* "END-of-entry-condition" ? */
#endif /* ?FUNZIP */
break;
if (G.dstrm.avail_in == 0) {
if (fillinbuf(__G) == 0) {
/* no "END-condition" yet, but no more data */
retval = 2; goto uzinflate_cleanup_exit;
}
G.dstrm.next_in = G.inptr;
G.dstrm.avail_in = G.incnt;
}
Trace((stderr, " avail_in = %u\n", G.dstrm.avail_in));
}
/* flush slide[] */
if ((retval = FLUSH(wsize - G.dstrm.avail_out)) != 0)
goto uzinflate_cleanup_exit;
Trace((stderr, "inside loop: flushing %ld bytes (ptr diff = %ld)\n",
(long)(wsize - G.dstrm.avail_out),
(long)(G.dstrm.next_out-(Bytef *)redirSlide)));
G.dstrm.next_out = redirSlide;
G.dstrm.avail_out = wsize;
}
/* no more input, so loop until we have all output */
Trace((stderr, "beginning final loop: err = %d\n", err));
repeated_buf_err = FALSE;
while (err != Z_STREAM_END) {
err = inflate(&G.dstrm, Z_PARTIAL_FLUSH);
if (err == Z_DATA_ERROR) {
retval = 2; goto uzinflate_cleanup_exit;
} else if (err == Z_MEM_ERROR) {
retval = 3; goto uzinflate_cleanup_exit;
} else if (err == Z_BUF_ERROR) { /* DEBUG */
#ifdef FUNZIP
Trace((stderr,
"zlib inflate() did not detect stream end\n"));
#else
Trace((stderr,
"zlib inflate() did not detect stream end (%s, %s)\n",
G.zipfn, G.filename));
#endif
if ((!repeated_buf_err) && (G.dstrm.avail_in == 0)) {
/* when detecting this problem for the first time,
try to provide one fake byte beyond "EOF"... */
G.dstrm.next_in = "";
G.dstrm.avail_in = 1;
repeated_buf_err = TRUE;
} else
break;
} else if (err != Z_OK && err != Z_STREAM_END) {
Trace((stderr, "oops! (inflate(final loop) err = %d)\n", err));
DESTROYGLOBALS();
EXIT(PK_MEM3);
}
/* final flush of slide[] */
if ((retval = FLUSH(wsize - G.dstrm.avail_out)) != 0)
goto uzinflate_cleanup_exit;
Trace((stderr, "final loop: flushing %ld bytes (ptr diff = %ld)\n",
(long)(wsize - G.dstrm.avail_out),
(long)(G.dstrm.next_out-(Bytef *)redirSlide)));
G.dstrm.next_out = redirSlide;
G.dstrm.avail_out = wsize;
}
Trace((stderr, "total in = %lu, total out = %lu\n", G.dstrm.total_in,
G.dstrm.total_out));
G.inptr = (uch *)G.dstrm.next_in;
G.incnt = (G.inbuf + INBUFSIZ) - G.inptr; /* reset for other routines */
uzinflate_cleanup_exit:
err = inflateReset(&G.dstrm);
if (err != Z_OK)
Trace((stderr, "oops! (inflateReset() err = %d)\n", err));
#endif /* ?USE_ZLIB_INFLATCB */
return retval;
}
/*---------------------------------------------------------------------------*/
#else /* !USE_ZLIB */
/* Function prototypes */
#ifndef OF
# ifdef __STDC__
# define OF(a) a
# else
# define OF(a) ()
# endif
#endif /* !OF */
int inflate_codes OF((__GPRO__ struct huft *tl, struct huft *td,
unsigned bl, unsigned bd));
static int inflate_stored OF((__GPRO));
static int inflate_fixed OF((__GPRO));
static int inflate_dynamic OF((__GPRO));
static int inflate_block OF((__GPRO__ int *e));
/* The inflate algorithm uses a sliding 32K byte window on the uncompressed
stream to find repeated byte strings. This is implemented here as a
circular buffer. The index is updated simply by incrementing and then
and'ing with 0x7fff (32K-1). */
/* It is left to other modules to supply the 32K area. It is assumed
to be usable as if it were declared "uch slide[32768];" or as just
"uch *slide;" and then malloc'ed in the latter case. The definition
must be in unzip.h, included above. */
/* unsigned wp; moved to globals.h */ /* current position in slide */
/* Tables for deflate from PKZIP's appnote.txt. */
/* - Order of the bit length code lengths */
static ZCONST unsigned border[] = {
16, 17, 18, 0, 8, 7, 9, 6, 10, 5, 11, 4, 12, 3, 13, 2, 14, 1, 15};
/* - Copy lengths for literal codes 257..285 */
#ifdef USE_DEFLATE64
static ZCONST ush cplens64[] = {
3, 4, 5, 6, 7, 8, 9, 10, 11, 13, 15, 17, 19, 23, 27, 31,
35, 43, 51, 59, 67, 83, 99, 115, 131, 163, 195, 227, 3, 0, 0};
/* For Deflate64, the code 285 is defined differently. */
#else
# define cplens32 cplens
#endif
static ZCONST ush cplens32[] = {
3, 4, 5, 6, 7, 8, 9, 10, 11, 13, 15, 17, 19, 23, 27, 31,
35, 43, 51, 59, 67, 83, 99, 115, 131, 163, 195, 227, 258, 0, 0};
/* note: see note #13 above about the 258 in this list. */
/* - Extra bits for literal codes 257..285 */
#ifdef USE_DEFLATE64
static ZCONST uch cplext64[] = {
0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2,
3, 3, 3, 3, 4, 4, 4, 4, 5, 5, 5, 5, 16, INVALID_CODE, INVALID_CODE};
#else
# define cplext32 cplext
#endif
static ZCONST uch cplext32[] = {
0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2,
3, 3, 3, 3, 4, 4, 4, 4, 5, 5, 5, 5, 0, INVALID_CODE, INVALID_CODE};
/* - Copy offsets for distance codes 0..29 (0..31 for Deflate64) */
static ZCONST ush cpdist[] = {
1, 2, 3, 4, 5, 7, 9, 13, 17, 25, 33, 49, 65, 97, 129, 193,
257, 385, 513, 769, 1025, 1537, 2049, 3073, 4097, 6145,
#if (defined(USE_DEFLATE64) || defined(PKZIP_BUG_WORKAROUND))
8193, 12289, 16385, 24577, 32769, 49153};
#else
8193, 12289, 16385, 24577};
#endif
/* - Extra bits for distance codes 0..29 (0..31 for Deflate64) */
#ifdef USE_DEFLATE64
static ZCONST uch cpdext64[] = {
0, 0, 0, 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6,
7, 7, 8, 8, 9, 9, 10, 10, 11, 11,
12, 12, 13, 13, 14, 14};
#else
# define cpdext32 cpdext
#endif
static ZCONST uch cpdext32[] = {
0, 0, 0, 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6,
7, 7, 8, 8, 9, 9, 10, 10, 11, 11,
#ifdef PKZIP_BUG_WORKAROUND
12, 12, 13, 13, INVALID_CODE, INVALID_CODE};
#else
12, 12, 13, 13};
#endif
#ifdef PKZIP_BUG_WORKAROUND
# define MAXLITLENS 288
#else
# define MAXLITLENS 286
#endif
#if (defined(USE_DEFLATE64) || defined(PKZIP_BUG_WORKAROUND))
# define MAXDISTS 32
#else
# define MAXDISTS 30
#endif
/* moved to consts.h (included in unzip.c), resp. funzip.c */
#if 0
/* And'ing with mask_bits[n] masks the lower n bits */
ZCONST unsigned near mask_bits[17] = {
0x0000,
0x0001, 0x0003, 0x0007, 0x000f, 0x001f, 0x003f, 0x007f, 0x00ff,
0x01ff, 0x03ff, 0x07ff, 0x0fff, 0x1fff, 0x3fff, 0x7fff, 0xffff
};
#endif /* 0 */
/* Macros for inflate() bit peeking and grabbing.
The usage is:
NEEDBITS(j)
x = b & mask_bits[j];
DUMPBITS(j)
where NEEDBITS makes sure that b has at least j bits in it, and
DUMPBITS removes the bits from b. The macros use the variable k
for the number of bits in b. Normally, b and k are register
variables for speed and are initialized at the beginning of a
routine that uses these macros from a global bit buffer and count.
In order to not ask for more bits than there are in the compressed
stream, the Huffman tables are constructed to only ask for just
enough bits to make up the end-of-block code (value 256). Then no
bytes need to be "returned" to the buffer at the end of the last
block. See the huft_build() routine.
Actually, the precautions mentioned above are not sufficient to
prevent fetches of bits beyound the end of the last block in every
case. When the last code fetched before the end-of-block code was
a very short distance code (shorter than "distance-prefetch-bits" -
"end-of-block code bits"), this last distance code fetch already
exausts the available data. To prevent failure of extraction in this
case, the "read beyond EOF" check delays the raise of the "invalid
data" error until an actual overflow of "used data" is detected.
This error condition is only fulfilled when the "number of available
bits" counter k is found to be negative in the NEEDBITS() macro.
An alternate fix for that problem adjusts the size of the distance code
base table so that it does not exceed the length of the end-of-block code
plus the minimum length of a distance code. This alternate fix can be
enabled by defining the preprocessor symbol FIX_PAST_EOB_BY_TABLEADJUST.
*/
/* These have been moved to globals.h */
#if 0
ulg bb; /* bit buffer */
unsigned bk; /* bits in bit buffer */
#endif
#ifndef CHECK_EOF
# define CHECK_EOF /* default as of 5.13/5.2 */
#endif
#ifndef CHECK_EOF
# define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE)<<k;k+=8;}}
#else
# ifdef FIX_PAST_EOB_BY_TABLEADJUST
# define NEEDBITS(n) {while(k<(n)){int c=NEXTBYTE;\
if(c==EOF){retval=1;goto cleanup_and_exit;}\
b|=((ulg)c)<<k;k+=8;}}
# else
# define NEEDBITS(n) {while((int)k<(int)(n)){int c=NEXTBYTE;\
if(c==EOF){if((int)k>=0)break;retval=1;goto cleanup_and_exit;}\
b|=((ulg)c)<<k;k+=8;}}
# endif
#endif
#define DUMPBITS(n) {b>>=(n);k-=(n);}
/*
Huffman code decoding is performed using a multi-level table lookup.
The fastest way to decode is to simply build a lookup table whose
size is determined by the longest code. However, the time it takes
to build this table can also be a factor if the data being decoded
are not very long. The most common codes are necessarily the
shortest codes, so those codes dominate the decoding time, and hence
the speed. The idea is you can have a shorter table that decodes the
shorter, more probable codes, and then point to subsidiary tables for
the longer codes. The time it costs to decode the longer codes is
then traded against the time it takes to make longer tables.
This results of this trade are in the variables lbits and dbits
below. lbits is the number of bits the first level table for literal/
length codes can decode in one step, and dbits is the same thing for
the distance codes. Subsequent tables are also less than or equal to
those sizes. These values may be adjusted either when all of the
codes are shorter than that, in which case the longest code length in
bits is used, or when the shortest code is *longer* than the requested
table size, in which case the length of the shortest code in bits is
used.
There are two different values for the two tables, since they code a
different number of possibilities each. The literal/length table
codes 286 possible values, or in a flat code, a little over eight
bits. The distance table codes 30 possible values, or a little less
than five bits, flat. The optimum values for speed end up being
about one bit more than those, so lbits is 8+1 and dbits is 5+1.
The optimum values may differ though from machine to machine, and
possibly even between compilers. Your mileage may vary.
*/
/* bits in base literal/length lookup table */
static ZCONST unsigned lbits = 9;
/* bits in base distance lookup table */
static ZCONST unsigned dbits = 6;
#ifndef ASM_INFLATECODES
int inflate_codes(__G__ tl, td, bl, bd)
__GDEF
struct huft *tl, *td; /* literal/length and distance decoder tables */
unsigned bl, bd; /* number of bits decoded by tl[] and td[] */
/* inflate (decompress) the codes in a deflated (compressed) block.
Return an error code or zero if it all goes ok. */
{
register unsigned e; /* table entry flag/number of extra bits */
unsigned d; /* index for copy */
UINT_D64 n; /* length for copy (deflate64: might be 64k+2) */
UINT_D64 w; /* current window position (deflate64: up to 64k) */
struct huft *t; /* pointer to table entry */
unsigned ml, md; /* masks for bl and bd bits */
register ulg b; /* bit buffer */
register unsigned k; /* number of bits in bit buffer */
int retval = 0; /* error code returned: initialized to "no error" */
/* make local copies of globals */
b = G.bb; /* initialize bit buffer */
k = G.bk;
w = G.wp; /* initialize window position */
/* inflate the coded data */
ml = mask_bits[bl]; /* precompute masks for speed */
md = mask_bits[bd];
while (1) /* do until end of block */
{
NEEDBITS(bl)
t = tl + ((unsigned)b & ml);
while (1) {
DUMPBITS(t->b)
if ((e = t->e) == 32) /* then it's a literal */
{
redirSlide[w++] = (uch)t->v.n;
if (w == wsize)
{
if ((retval = FLUSH(w)) != 0) goto cleanup_and_exit;
w = 0;
}
break;
}
if (e < 31) /* then it's a length */
{
/* get length of block to copy */
NEEDBITS(e)
n = t->v.n + ((unsigned)b & mask_bits[e]);
DUMPBITS(e)
/* decode distance of block to copy */
NEEDBITS(bd)
t = td + ((unsigned)b & md);
while (1) {
DUMPBITS(t->b)
if ((e = t->e) < 32)
break;
if (IS_INVALID_CODE(e))
return 1;
e &= 31;
NEEDBITS(e)
t = t->v.t + ((unsigned)b & mask_bits[e]);
}
NEEDBITS(e)
d = (unsigned)w - t->v.n - ((unsigned)b & mask_bits[e]);
DUMPBITS(e)
/* do the copy */
do {
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
if (G.redirect_slide) {
/* &= w/ wsize unnecessary & wrong if redirect */
if ((UINT_D64)d >= wsize)
return 1; /* invalid compressed data */
e = (unsigned)(wsize - (d > (unsigned)w ? (UINT_D64)d : w));
}
else
#endif
e = (unsigned)(wsize -
((d &= (unsigned)(wsize-1)) > (unsigned)w ?
(UINT_D64)d : w));
if ((UINT_D64)e > n) e = (unsigned)n;
n -= e;
#ifndef NOMEMCPY
if ((unsigned)w - d >= e)
/* (this test assumes unsigned comparison) */
{
memcpy(redirSlide + (unsigned)w, redirSlide + d, e);
w += e;
d += e;
}
else /* do it slowly to avoid memcpy() overlap */
#endif /* !NOMEMCPY */
do {
redirSlide[w++] = redirSlide[d++];
} while (--e);
if (w == wsize)
{
if ((retval = FLUSH(w)) != 0) goto cleanup_and_exit;
w = 0;
}
} while (n);
break;
}
if (e == 31) /* it's the EOB signal */
{
/* sorry for this goto, but we have to exit two loops at once */
goto cleanup_decode;
}
if (IS_INVALID_CODE(e))
return 1;
e &= 31;
NEEDBITS(e)
t = t->v.t + ((unsigned)b & mask_bits[e]);
}
}
cleanup_decode:
/* restore the globals from the locals */
G.wp = (unsigned)w; /* restore global window pointer */
G.bb = b; /* restore global bit buffer */
G.bk = k;
cleanup_and_exit:
/* done */
return retval;
}
#endif /* ASM_INFLATECODES */
static int inflate_stored(__G)
__GDEF
/* "decompress" an inflated type 0 (stored) block. */
{
UINT_D64 w; /* current window position (deflate64: up to 64k!) */
unsigned n; /* number of bytes in block */
register ulg b; /* bit buffer */
register unsigned k; /* number of bits in bit buffer */
int retval = 0; /* error code returned: initialized to "no error" */
/* make local copies of globals */
Trace((stderr, "\nstored block"));
b = G.bb; /* initialize bit buffer */
k = G.bk;
w = G.wp; /* initialize window position */
/* go to byte boundary */
n = k & 7;
DUMPBITS(n);
/* get the length and its complement */
NEEDBITS(16)
n = ((unsigned)b & 0xffff);
DUMPBITS(16)
NEEDBITS(16)
if (n != (unsigned)((~b) & 0xffff))
return 1; /* error in compressed data */
DUMPBITS(16)
/* read and output the compressed data */
while (n--)
{
NEEDBITS(8)
redirSlide[w++] = (uch)b;
if (w == wsize)
{
if ((retval = FLUSH(w)) != 0) goto cleanup_and_exit;
w = 0;
}
DUMPBITS(8)
}
/* restore the globals from the locals */
G.wp = (unsigned)w; /* restore global window pointer */
G.bb = b; /* restore global bit buffer */
G.bk = k;
cleanup_and_exit:
return retval;
}
/* Globals for literal tables (built once) */
/* Moved to globals.h */
#if 0
struct huft *fixed_tl = (struct huft *)NULL;
struct huft *fixed_td;
int fixed_bl, fixed_bd;
#endif
static int inflate_fixed(__G)
__GDEF
/* decompress an inflated type 1 (fixed Huffman codes) block. We should
either replace this with a custom decoder, or at least precompute the
Huffman tables. */
{
/* if first time, set up tables for fixed blocks */
Trace((stderr, "\nliteral block"));
if (G.fixed_tl == (struct huft *)NULL)
{
int i; /* temporary variable */
unsigned l[288]; /* length list for huft_build */
/* literal table */
for (i = 0; i < 144; i++)
l[i] = 8;
for (; i < 256; i++)
l[i] = 9;
for (; i < 280; i++)
l[i] = 7;
for (; i < 288; i++) /* make a complete, but wrong code set */
l[i] = 8;
G.fixed_bl = 7;
#ifdef USE_DEFLATE64
if ((i = huft_build(__G__ l, 288, 257, G.cplens, G.cplext,
&G.fixed_tl, &G.fixed_bl)) != 0)
#else
if ((i = huft_build(__G__ l, 288, 257, cplens, cplext,
&G.fixed_tl, &G.fixed_bl)) != 0)
#endif
{
G.fixed_tl = (struct huft *)NULL;
return i;
}
/* distance table */
for (i = 0; i < MAXDISTS; i++) /* make an incomplete code set */
l[i] = 5;
G.fixed_bd = 5;
#ifdef USE_DEFLATE64
if ((i = huft_build(__G__ l, MAXDISTS, 0, cpdist, G.cpdext,
&G.fixed_td, &G.fixed_bd)) > 1)
#else
if ((i = huft_build(__G__ l, MAXDISTS, 0, cpdist, cpdext,
&G.fixed_td, &G.fixed_bd)) > 1)
#endif
{
huft_free(G.fixed_tl);
G.fixed_td = G.fixed_tl = (struct huft *)NULL;
return i;
}
}
/* decompress until an end-of-block code */
return inflate_codes(__G__ G.fixed_tl, G.fixed_td,
G.fixed_bl, G.fixed_bd);
}
static int inflate_dynamic(__G)
__GDEF
/* decompress an inflated type 2 (dynamic Huffman codes) block. */
{
unsigned i; /* temporary variables */
unsigned j;
unsigned l; /* last length */
unsigned m; /* mask for bit lengths table */
unsigned n; /* number of lengths to get */
struct huft *tl = (struct huft *)NULL; /* literal/length code table */
struct huft *td = (struct huft *)NULL; /* distance code table */
struct huft *th; /* temp huft table pointer used in tables decoding */
unsigned bl; /* lookup bits for tl */
unsigned bd; /* lookup bits for td */
unsigned nb; /* number of bit length codes */
unsigned nl; /* number of literal/length codes */
unsigned nd; /* number of distance codes */
unsigned ll[MAXLITLENS+MAXDISTS]; /* lit./length and distance code lengths */
register ulg b; /* bit buffer */
register unsigned k; /* number of bits in bit buffer */
int retval = 0; /* error code returned: initialized to "no error" */
/* make local bit buffer */
Trace((stderr, "\ndynamic block"));
b = G.bb;
k = G.bk;
/* read in table lengths */
NEEDBITS(5)
nl = 257 + ((unsigned)b & 0x1f); /* number of literal/length codes */
DUMPBITS(5)
NEEDBITS(5)
nd = 1 + ((unsigned)b & 0x1f); /* number of distance codes */
DUMPBITS(5)
NEEDBITS(4)
nb = 4 + ((unsigned)b & 0xf); /* number of bit length codes */
DUMPBITS(4)
if (nl > MAXLITLENS || nd > MAXDISTS)
return 1; /* bad lengths */
/* read in bit-length-code lengths */
for (j = 0; j < nb; j++)
{
NEEDBITS(3)
ll[border[j]] = (unsigned)b & 7;
DUMPBITS(3)
}
for (; j < 19; j++)
ll[border[j]] = 0;
/* build decoding table for trees--single level, 7 bit lookup */
bl = 7;
retval = huft_build(__G__ ll, 19, 19, NULL, NULL, &tl, &bl);
if (bl == 0) /* no bit lengths */
retval = 1;
if (retval)
{
if (retval == 1)
huft_free(tl);
return retval; /* incomplete code set */
}
/* read in literal and distance code lengths */
n = nl + nd;
m = mask_bits[bl];
i = l = 0;
while (i < n)
{
NEEDBITS(bl)
j = (th = tl + ((unsigned)b & m))->b;
DUMPBITS(j)
j = th->v.n;
if (j < 16) /* length of code in bits (0..15) */
ll[i++] = l = j; /* save last length in l */
else if (j == 16) /* repeat last length 3 to 6 times */
{
NEEDBITS(2)
j = 3 + ((unsigned)b & 3);
DUMPBITS(2)
if ((unsigned)i + j > n) {
huft_free(tl);
return 1;
}
while (j--)
ll[i++] = l;
}
else if (j == 17) /* 3 to 10 zero length codes */
{
NEEDBITS(3)
j = 3 + ((unsigned)b & 7);
DUMPBITS(3)
if ((unsigned)i + j > n) {
huft_free(tl);
return 1;
}
while (j--)
ll[i++] = 0;
l = 0;
}
else /* j == 18: 11 to 138 zero length codes */
{
NEEDBITS(7)
j = 11 + ((unsigned)b & 0x7f);
DUMPBITS(7)
if ((unsigned)i + j > n) {
huft_free(tl);
return 1;
}
while (j--)
ll[i++] = 0;
l = 0;
}
}
/* free decoding table for trees */
huft_free(tl);
/* restore the global bit buffer */
G.bb = b;
G.bk = k;
/* build the decoding tables for literal/length and distance codes */
bl = lbits;
#ifdef USE_DEFLATE64
retval = huft_build(__G__ ll, nl, 257, G.cplens, G.cplext, &tl, &bl);
#else
retval = huft_build(__G__ ll, nl, 257, cplens, cplext, &tl, &bl);
#endif
if (bl == 0) /* no literals or lengths */
retval = 1;
if (retval)
{
if (retval == 1) {
if (!uO.qflag)
MESSAGE((uch *)"(incomplete l-tree) ", 21L, 1);
huft_free(tl);
}
return retval; /* incomplete code set */
}
#ifdef FIX_PAST_EOB_BY_TABLEADJUST
/* Adjust the requested distance base table size so that a distance code
fetch never tries to get bits behind an immediatly following end-of-block
code. */
bd = (dbits <= bl+1 ? dbits : bl+1);
#else
bd = dbits;
#endif
#ifdef USE_DEFLATE64
retval = huft_build(__G__ ll + nl, nd, 0, cpdist, G.cpdext, &td, &bd);
#else
retval = huft_build(__G__ ll + nl, nd, 0, cpdist, cpdext, &td, &bd);
#endif
#ifdef PKZIP_BUG_WORKAROUND
if (retval == 1)
retval = 0;
#endif
if (bd == 0 && nl > 257) /* lengths but no distances */
retval = 1;
if (retval)
{
if (retval == 1) {
if (!uO.qflag)
MESSAGE((uch *)"(incomplete d-tree) ", 21L, 1);
huft_free(td);
}
huft_free(tl);
return retval;
}
/* decompress until an end-of-block code */
retval = inflate_codes(__G__ tl, td, bl, bd);
cleanup_and_exit:
/* free the decoding tables, return */
if (tl != (struct huft *)NULL)
huft_free(tl);
if (td != (struct huft *)NULL)
huft_free(td);
return retval;
}
static int inflate_block(__G__ e)
__GDEF
int *e; /* last block flag */
/* decompress an inflated block */
{
unsigned t; /* block type */
register ulg b; /* bit buffer */
register unsigned k; /* number of bits in bit buffer */
int retval = 0; /* error code returned: initialized to "no error" */
/* make local bit buffer */
b = G.bb;
k = G.bk;
/* read in last block bit */
NEEDBITS(1)
*e = (int)b & 1;
DUMPBITS(1)
/* read in block type */
NEEDBITS(2)
t = (unsigned)b & 3;
DUMPBITS(2)
/* restore the global bit buffer */
G.bb = b;
G.bk = k;
/* inflate that block type */
if (t == 2)
return inflate_dynamic(__G);
if (t == 0)
return inflate_stored(__G);
if (t == 1)
return inflate_fixed(__G);
/* bad block type */
retval = 2;
cleanup_and_exit:
return retval;
}
int inflate(__G__ is_defl64)
__GDEF
int is_defl64;
/* decompress an inflated entry */
{
int e; /* last block flag */
int r; /* result code */
#ifdef DEBUG
unsigned h = 0; /* maximum struct huft's malloc'ed */
#endif
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
if (G.redirect_slide)
wsize = G.redirect_size, redirSlide = G.redirect_buffer;
else
wsize = WSIZE, redirSlide = slide; /* how they're #defined if !DLL */
#endif
/* initialize window, bit buffer */
G.wp = 0;
G.bk = 0;
G.bb = 0;
#ifdef USE_DEFLATE64
if (is_defl64) {
G.cplens = cplens64;
G.cplext = cplext64;
G.cpdext = cpdext64;
G.fixed_tl = G.fixed_tl64;
G.fixed_bl = G.fixed_bl64;
G.fixed_td = G.fixed_td64;
G.fixed_bd = G.fixed_bd64;
} else {
G.cplens = cplens32;
G.cplext = cplext32;
G.cpdext = cpdext32;
G.fixed_tl = G.fixed_tl32;
G.fixed_bl = G.fixed_bl32;
G.fixed_td = G.fixed_td32;
G.fixed_bd = G.fixed_bd32;
}
#else /* !USE_DEFLATE64 */
if (is_defl64) {
/* This should not happen unless UnZip is built from object files
* compiled with inconsistent option setting. Handle this by
* returning with "bad input" error code.
*/
Trace((stderr, "\nThis inflate() cannot handle Deflate64!\n"));
return 2;
}
#endif /* ?USE_DEFLATE64 */
/* decompress until the last block */
do {
#ifdef DEBUG
G.hufts = 0;
#endif
if ((r = inflate_block(__G__ &e)) != 0)
return r;
#ifdef DEBUG
if (G.hufts > h)
h = G.hufts;
#endif
} while (!e);
Trace((stderr, "\n%u bytes in Huffman tables (%u/entry)\n",
h * (unsigned)sizeof(struct huft), (unsigned)sizeof(struct huft)));
#ifdef USE_DEFLATE64
if (is_defl64) {
G.fixed_tl64 = G.fixed_tl;
G.fixed_bl64 = G.fixed_bl;
G.fixed_td64 = G.fixed_td;
G.fixed_bd64 = G.fixed_bd;
} else {
G.fixed_tl32 = G.fixed_tl;
G.fixed_bl32 = G.fixed_bl;
G.fixed_td32 = G.fixed_td;
G.fixed_bd32 = G.fixed_bd;
}
#endif
/* flush out redirSlide and return (success, unless final FLUSH failed) */
return (FLUSH(G.wp));
}
int inflate_free(__G)
__GDEF
{
if (G.fixed_tl != (struct huft *)NULL)
{
huft_free(G.fixed_td);
huft_free(G.fixed_tl);
G.fixed_td = G.fixed_tl = (struct huft *)NULL;
}
return 0;
}
#endif /* ?USE_ZLIB */
/*
* GRR: moved huft_build() and huft_free() down here; used by explode()
* and fUnZip regardless of whether USE_ZLIB defined or not
*/
/* If BMAX needs to be larger than 16, then h and x[] should be ulg. */
#define BMAX 16 /* maximum bit length of any code (16 for explode) */
#define N_MAX 288 /* maximum number of codes in any set */
int huft_build(__G__ b, n, s, d, e, t, m)
__GDEF
ZCONST unsigned *b; /* code lengths in bits (all assumed <= BMAX) */
unsigned n; /* number of codes (assumed <= N_MAX) */
unsigned s; /* number of simple-valued codes (0..s-1) */
ZCONST ush *d; /* list of base values for non-simple codes */
ZCONST uch *e; /* list of extra bits for non-simple codes */
struct huft **t; /* result: starting table */
unsigned *m; /* maximum lookup bits, returns actual */
/* Given a list of code lengths and a maximum table size, make a set of
tables to decode that set of codes. Return zero on success, one if
the given code set is incomplete (the tables are still built in this
case), two if the input is invalid (all zero length codes or an
oversubscribed set of lengths), and three if not enough memory.
The code with value 256 is special, and the tables are constructed
so that no bits beyond that code are fetched when that code is
decoded. */
{
unsigned a; /* counter for codes of length k */
unsigned c[BMAX+1]; /* bit length count table */
unsigned el; /* length of EOB code (value 256) */
unsigned f; /* i repeats in table every f entries */
int g; /* maximum code length */
int h; /* table level */
register unsigned i; /* counter, current code */
register unsigned j; /* counter */
register int k; /* number of bits in current code */
int lx[BMAX+1]; /* memory for l[-1..BMAX-1] */
int *l = lx+1; /* stack of bits per table */
register unsigned *p; /* pointer into c[], b[], or v[] */
register struct huft *q; /* points to current table */
struct huft r; /* table entry for structure assignment */
struct huft *u[BMAX]; /* table stack */
unsigned v[N_MAX]; /* values in order of bit length */
register int w; /* bits before this table == (l * h) */
unsigned x[BMAX+1]; /* bit offsets, then code stack */
unsigned *xp; /* pointer into x */
int y; /* number of dummy codes added */
unsigned z; /* number of entries in current table */
/* Generate counts for each bit length */
el = n > 256 ? b[256] : BMAX; /* set length of EOB code, if any */
memzero((char *)c, sizeof(c));
p = (unsigned *)b; i = n;
do {
c[*p]++; p++; /* assume all entries <= BMAX */
} while (--i);
if (c[0] == n) /* null input--all zero length codes */
{
*t = (struct huft *)NULL;
*m = 0;
return 0;
}
/* Find minimum and maximum length, bound *m by those */
for (j = 1; j <= BMAX; j++)
if (c[j])
break;
k = j; /* minimum code length */
if (*m < j)
*m = j;
for (i = BMAX; i; i--)
if (c[i])
break;
g = i; /* maximum code length */
if (*m > i)
*m = i;
/* Adjust last length count to fill out codes, if needed */
for (y = 1 << j; j < i; j++, y <<= 1)
if ((y -= c[j]) < 0)
return 2; /* bad input: more codes than bits */
if ((y -= c[i]) < 0)
return 2;
c[i] += y;
/* Generate starting offsets into the value table for each length */
x[1] = j = 0;
p = c + 1; xp = x + 2;
while (--i) { /* note that i == g from above */
*xp++ = (j += *p++);
}
/* Make a table of values in order of bit lengths */
memzero((char *)v, sizeof(v));
p = (unsigned *)b; i = 0;
do {
if ((j = *p++) != 0)
v[x[j]++] = i;
} while (++i < n);
n = x[g]; /* set n to length of v */
/* Generate the Huffman codes and for each, make the table entries */
x[0] = i = 0; /* first Huffman code is zero */
p = v; /* grab values in bit order */
h = -1; /* no tables yet--level -1 */
w = l[-1] = 0; /* no bits decoded yet */
u[0] = (struct huft *)NULL; /* just to keep compilers happy */
q = (struct huft *)NULL; /* ditto */
z = 0; /* ditto */
/* go through the bit lengths (k already is bits in shortest code) */
for (; k <= g; k++)
{
a = c[k];
while (a--)
{
/* here i is the Huffman code of length k bits for value *p */
/* make tables up to required level */
while (k > w + l[h])
{
w += l[h++]; /* add bits already decoded */
/* compute minimum size table less than or equal to *m bits */
z = (z = g - w) > *m ? *m : z; /* upper limit */
if ((f = 1 << (j = k - w)) > a + 1) /* try a k-w bit table */
{ /* too few codes for k-w bit table */
f -= a + 1; /* deduct codes from patterns left */
xp = c + k;
while (++j < z) /* try smaller tables up to z bits */
{
if ((f <<= 1) <= *++xp)
break; /* enough codes to use up j bits */
f -= *xp; /* else deduct codes from patterns */
}
}
if ((unsigned)w + j > el && (unsigned)w < el)
j = el - w; /* make EOB code end at table */
z = 1 << j; /* table entries for j-bit table */
l[h] = j; /* set table size in stack */
/* allocate and link in new table */
if ((q = (struct huft *)malloc((z + 1)*sizeof(struct huft))) ==
(struct huft *)NULL)
{
if (h)
huft_free(u[0]);
return 3; /* not enough memory */
}
#ifdef DEBUG
G.hufts += z + 1; /* track memory usage */
#endif
*t = q + 1; /* link to list for huft_free() */
*(t = &(q->v.t)) = (struct huft *)NULL;
u[h] = ++q; /* table starts after link */
/* connect to last table, if there is one */
if (h)
{
x[h] = i; /* save pattern for backing up */
r.b = (uch)l[h-1]; /* bits to dump before this table */
r.e = (uch)(32 + j); /* bits in this table */
r.v.t = q; /* pointer to this table */
j = (i & ((1 << w) - 1)) >> (w - l[h-1]);
u[h-1][j] = r; /* connect to last table */
}
}
/* set up table entry in r */
r.b = (uch)(k - w);
if (p >= v + n)
r.e = INVALID_CODE; /* out of values--invalid code */
else if (*p < s)
{
r.e = (uch)(*p < 256 ? 32 : 31); /* 256 is end-of-block code */
r.v.n = (ush)*p++; /* simple code is just the value */
}
else
{
r.e = e[*p - s]; /* non-simple--look up in lists */
r.v.n = d[*p++ - s];
}
/* fill code-like entries with r */
f = 1 << (k - w);
for (j = i >> w; j < z; j += f)
q[j] = r;
/* backwards increment the k-bit code i */
for (j = 1 << (k - 1); i & j; j >>= 1)
i ^= j;
i ^= j;
/* backup over finished tables */
while ((i & ((1 << w) - 1)) != x[h])
w -= l[--h]; /* don't need to update q */
}
}
/* return actual size of base table */
*m = l[0];
/* Return true (1) if we were given an incomplete table */
return y != 0 && g != 1;
}
int huft_free(t)
struct huft *t; /* table to free */
/* Free the malloc'ed tables built by huft_build(), which makes a linked
list of the tables it made, with the links in a dummy first entry of
each table. */
{
register struct huft *p, *q;
/* Go through linked list, freeing from the malloced (t[-1]) address. */
p = t;
while (p != (struct huft *)NULL)
{
q = (--p)->v.t;
free((zvoid *)p);
p = q;
}
return 0;
}
|