which included commits to RCS files with non-trunk default branches. git-svn-id: svn://10.65.10.50/trunk@5403 c028cbd2-c16b-5b4b-a496-9718f37d4682
		
			
				
	
	
		
			857 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			C
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			857 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			C
		
	
	
		
			Executable File
		
	
	
	
	
| /* explode.c -- put in the public domain by Mark Adler
 | |
|    version c15, 6 July 1996 */
 | |
| 
 | |
| 
 | |
| /* You can do whatever you like with this source file, though I would
 | |
|    prefer that if you modify it and redistribute it that you include
 | |
|    comments to that effect with your name and the date.  Thank you.
 | |
| 
 | |
|    History:
 | |
|    vers    date          who           what
 | |
|    ----  ---------  --------------  ------------------------------------
 | |
|     c1   30 Mar 92  M. Adler        explode that uses huft_build from inflate
 | |
|                                     (this gives over a 70% speed improvement
 | |
|                                     over the original unimplode.c, which
 | |
|                                     decoded a bit at a time)
 | |
|     c2    4 Apr 92  M. Adler        fixed bug for file sizes a multiple of 32k.
 | |
|     c3   10 Apr 92  M. Adler        added a little memory tracking if DEBUG
 | |
|     c4   11 Apr 92  M. Adler        added NOMEMCPY do kill use of memcpy()
 | |
|     c5   21 Apr 92  M. Adler        added the WSIZE #define to allow reducing
 | |
|                                     the 32K window size for specialized
 | |
|                                     applications.
 | |
|     c6   31 May 92  M. Adler        added typecasts to eliminate some warnings
 | |
|     c7   27 Jun 92  G. Roelofs      added more typecasts.
 | |
|     c8   17 Oct 92  G. Roelofs      changed ULONG/UWORD/byte to ulg/ush/uch.
 | |
|     c9   19 Jul 93  J. Bush         added more typecasts (to return values);
 | |
|                                     made l[256] array static for Amiga.
 | |
|     c10   8 Oct 93  G. Roelofs      added used_csize for diagnostics; added
 | |
|                                     buf and unshrink arguments to flush();
 | |
|                                     undef'd various macros at end for Turbo C;
 | |
|                                     removed NEXTBYTE macro (now in unzip.h)
 | |
|                                     and bytebuf variable (not used); changed
 | |
|                                     memset() to memzero().
 | |
|     c11   9 Jan 94  M. Adler        fixed incorrect used_csize calculation.
 | |
|     c12   9 Apr 94  G. Roelofs      fixed split comments on preprocessor lines
 | |
|                                     to avoid bug in Encore compiler.
 | |
|     c13  25 Aug 94  M. Adler        fixed distance-length comment (orig c9 fix)
 | |
|     c14  22 Nov 95  S. Maxwell      removed unnecessary "static" on auto array
 | |
|     c15   6 Jul 96  W. Haidinger    added ulg typecasts to flush() calls
 | |
|  */
 | |
| 
 | |
| 
 | |
| /*
 | |
|    Explode imploded (PKZIP method 6 compressed) data.  This compression
 | |
|    method searches for as much of the current string of bytes (up to a length
 | |
|    of ~320) in the previous 4K or 8K bytes.  If it doesn't find any matches
 | |
|    (of at least length 2 or 3), it codes the next byte.  Otherwise, it codes
 | |
|    the length of the matched string and its distance backwards from the
 | |
|    current position.  Single bytes ("literals") are preceded by a one (a
 | |
|    single bit) and are either uncoded (the eight bits go directly into the
 | |
|    compressed stream for a total of nine bits) or Huffman coded with a
 | |
|    supplied literal code tree.  If literals are coded, then the minimum match
 | |
|    length is three, otherwise it is two.
 | |
| 
 | |
|    There are therefore four kinds of imploded streams: 8K search with coded
 | |
|    literals (min match = 3), 4K search with coded literals (min match = 3),
 | |
|    8K with uncoded literals (min match = 2), and 4K with uncoded literals
 | |
|    (min match = 2).  The kind of stream is identified in two bits of a
 | |
|    general purpose bit flag that is outside of the compressed stream.
 | |
| 
 | |
|    Distance-length pairs for matched strings are preceded by a zero bit (to
 | |
|    distinguish them from literals) and are always coded.  The distance comes
 | |
|    first and is either the low six (4K) or low seven (8K) bits of the
 | |
|    distance (uncoded), followed by the high six bits of the distance coded.
 | |
|    Then the length is six bits coded (0..63 + min match length), and if the
 | |
|    maximum such length is coded, then it's followed by another eight bits
 | |
|    (uncoded) to be added to the coded length.  This gives a match length
 | |
|    range of 2..320 or 3..321 bytes.
 | |
| 
 | |
|    The literal, length, and distance codes are all represented in a slightly
 | |
|    compressed form themselves.  What is sent are the lengths of the codes for
 | |
|    each value, which is sufficient to construct the codes.  Each byte of the
 | |
|    code representation is the code length (the low four bits representing
 | |
|    1..16), and the number of values sequentially with that length (the high
 | |
|    four bits also representing 1..16).  There are 256 literal code values (if
 | |
|    literals are coded), 64 length code values, and 64 distance code values,
 | |
|    in that order at the beginning of the compressed stream.  Each set of code
 | |
|    values is preceded (redundantly) with a byte indicating how many bytes are
 | |
|    in the code description that follows, in the range 1..256.
 | |
| 
 | |
|    The codes themselves are decoded using tables made by huft_build() from
 | |
|    the bit lengths.  That routine and its comments are in the inflate.c
 | |
|    module.
 | |
|  */
 | |
| 
 | |
| #define UNZIP_INTERNAL
 | |
| #include "unzip.h"      /* must supply slide[] (uch) array and NEXTBYTE macro */
 | |
| 
 | |
| #ifndef WSIZE
 | |
| #  define WSIZE 0x8000  /* window size--must be a power of two, and */
 | |
| #endif                  /* at least 8K for zip's implode method */
 | |
| 
 | |
| #ifdef DLL
 | |
| #  define wsize G._wsize
 | |
| #else
 | |
| #  define wsize WSIZE
 | |
| #endif
 | |
| 
 | |
| /* routines here */
 | |
| static int get_tree OF((__GPRO__ unsigned *l, unsigned n));
 | |
| static int explode_lit8 OF((__GPRO__ struct huft *tb, struct huft *tl,
 | |
|                             struct huft *td, int bb, int bl, int bd));
 | |
| static int explode_lit4 OF((__GPRO__ struct huft *tb, struct huft *tl,
 | |
|                             struct huft *td, int bb, int bl, int bd));
 | |
| static int explode_nolit8 OF((__GPRO__ struct huft *tl, struct huft *td,
 | |
|                               int bl, int bd));
 | |
| static int explode_nolit4 OF((__GPRO__ struct huft *tl, struct huft *td,
 | |
|                               int bl, int bd));
 | |
| int explode OF((__GPRO));
 | |
| 
 | |
| 
 | |
| /* The implode algorithm uses a sliding 4K or 8K byte window on the
 | |
|    uncompressed stream to find repeated byte strings.  This is implemented
 | |
|    here as a circular buffer.  The index is updated simply by incrementing
 | |
|    and then and'ing with 0x0fff (4K-1) or 0x1fff (8K-1).  Here, the 32K
 | |
|    buffer of inflate is used, and it works just as well to always have
 | |
|    a 32K circular buffer, so the index is anded with 0x7fff.  This is
 | |
|    done to allow the window to also be used as the output buffer. */
 | |
| /* This must be supplied in an external module useable like "uch slide[8192];"
 | |
|    or "uch *slide;", where the latter would be malloc'ed.  In unzip, slide[]
 | |
|    is actually a 32K area for use by inflate, which uses a 32K sliding window.
 | |
|  */
 | |
| 
 | |
| 
 | |
| /* Tables for length and distance */
 | |
| static ush cplen2[] =
 | |
|         {2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17,
 | |
|         18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34,
 | |
|         35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51,
 | |
|         52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65};
 | |
| static ush cplen3[] =
 | |
|         {3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18,
 | |
|         19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35,
 | |
|         36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52,
 | |
|         53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66};
 | |
| static ush extra[] =
 | |
|         {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 | |
|         0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 | |
|         0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 | |
|         8};
 | |
| static ush cpdist4[] =
 | |
|         {1, 65, 129, 193, 257, 321, 385, 449, 513, 577, 641, 705,
 | |
|         769, 833, 897, 961, 1025, 1089, 1153, 1217, 1281, 1345, 1409, 1473,
 | |
|         1537, 1601, 1665, 1729, 1793, 1857, 1921, 1985, 2049, 2113, 2177,
 | |
|         2241, 2305, 2369, 2433, 2497, 2561, 2625, 2689, 2753, 2817, 2881,
 | |
|         2945, 3009, 3073, 3137, 3201, 3265, 3329, 3393, 3457, 3521, 3585,
 | |
|         3649, 3713, 3777, 3841, 3905, 3969, 4033};
 | |
| static ush cpdist8[] =
 | |
|         {1, 129, 257, 385, 513, 641, 769, 897, 1025, 1153, 1281,
 | |
|         1409, 1537, 1665, 1793, 1921, 2049, 2177, 2305, 2433, 2561, 2689,
 | |
|         2817, 2945, 3073, 3201, 3329, 3457, 3585, 3713, 3841, 3969, 4097,
 | |
|         4225, 4353, 4481, 4609, 4737, 4865, 4993, 5121, 5249, 5377, 5505,
 | |
|         5633, 5761, 5889, 6017, 6145, 6273, 6401, 6529, 6657, 6785, 6913,
 | |
|         7041, 7169, 7297, 7425, 7553, 7681, 7809, 7937, 8065};
 | |
| 
 | |
| 
 | |
| /* Macros for inflate() bit peeking and grabbing.
 | |
|    The usage is:
 | |
| 
 | |
|         NEEDBITS(j)
 | |
|         x = b & mask_bits[j];
 | |
|         DUMPBITS(j)
 | |
| 
 | |
|    where NEEDBITS makes sure that b has at least j bits in it, and
 | |
|    DUMPBITS removes the bits from b.  The macros use the variable k
 | |
|    for the number of bits in b.  Normally, b and k are register
 | |
|    variables for speed.
 | |
|  */
 | |
| 
 | |
| #define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE)<<k;k+=8;}}
 | |
| #define DUMPBITS(n) {b>>=(n);k-=(n);}
 | |
| 
 | |
| 
 | |
| 
 | |
| static int get_tree(__G__ l, n)
 | |
|      __GDEF
 | |
| unsigned *l;            /* bit lengths */
 | |
| unsigned n;             /* number expected */
 | |
| /* Get the bit lengths for a code representation from the compressed
 | |
|    stream.  If get_tree() returns 4, then there is an error in the data.
 | |
|    Otherwise zero is returned. */
 | |
| {
 | |
|   unsigned i;           /* bytes remaining in list */
 | |
|   unsigned k;           /* lengths entered */
 | |
|   unsigned j;           /* number of codes */
 | |
|   unsigned b;           /* bit length for those codes */
 | |
| 
 | |
| 
 | |
|   /* get bit lengths */
 | |
|   i = NEXTBYTE + 1;                     /* length/count pairs to read */
 | |
|   k = 0;                                /* next code */
 | |
|   do {
 | |
|     b = ((j = NEXTBYTE) & 0xf) + 1;     /* bits in code (1..16) */
 | |
|     j = ((j & 0xf0) >> 4) + 1;          /* codes with those bits (1..16) */
 | |
|     if (k + j > n)
 | |
|       return 4;                         /* don't overflow l[] */
 | |
|     do {
 | |
|       l[k++] = b;
 | |
|     } while (--j);
 | |
|   } while (--i);
 | |
|   return k != n ? 4 : 0;                /* should have read n of them */
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| static int explode_lit8(__G__ tb, tl, td, bb, bl, bd)
 | |
|      __GDEF
 | |
| struct huft *tb, *tl, *td;      /* literal, length, and distance tables */
 | |
| int bb, bl, bd;                 /* number of bits decoded by those */
 | |
| /* Decompress the imploded data using coded literals and an 8K sliding
 | |
|    window. */
 | |
| {
 | |
|   long s;               /* bytes to decompress */
 | |
|   register unsigned e;  /* table entry flag/number of extra bits */
 | |
|   unsigned n, d;        /* length and index for copy */
 | |
|   unsigned w;           /* current window position */
 | |
|   struct huft *t;       /* pointer to table entry */
 | |
|   unsigned mb, ml, md;  /* masks for bb, bl, and bd bits */
 | |
|   register ulg b;       /* bit buffer */
 | |
|   register unsigned k;  /* number of bits in bit buffer */
 | |
|   unsigned u;           /* true if unflushed */
 | |
| 
 | |
| 
 | |
|   /* explode the coded data */
 | |
|   b = k = w = 0;                /* initialize bit buffer, window */
 | |
|   u = 1;                        /* buffer unflushed */
 | |
|   mb = mask_bits[bb];           /* precompute masks for speed */
 | |
|   ml = mask_bits[bl];
 | |
|   md = mask_bits[bd];
 | |
|   s = G.ucsize;
 | |
|   while (s > 0)                 /* do until ucsize bytes uncompressed */
 | |
|   {
 | |
|     NEEDBITS(1)
 | |
|     if (b & 1)                  /* then literal--decode it */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       s--;
 | |
|       NEEDBITS((unsigned)bb)    /* get coded literal */
 | |
|       if ((e = (t = tb + ((~(unsigned)b) & mb))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       redirSlide[w++] = (uch)t->v.n;
 | |
|       if (w == wsize)
 | |
|       {
 | |
|         flush(__G__ redirSlide, (ulg)w, 0);
 | |
|         w = u = 0;
 | |
|       }
 | |
|     }
 | |
|     else                        /* else distance/length */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       NEEDBITS(7)               /* get distance low bits */
 | |
|       d = (unsigned)b & 0x7f;
 | |
|       DUMPBITS(7)
 | |
|       NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | |
|       if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       d = w - d - t->v.n;       /* construct offset */
 | |
|       NEEDBITS((unsigned)bl)    /* get coded length */
 | |
|       if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       n = t->v.n;
 | |
|       if (e)                    /* get length extra bits */
 | |
|       {
 | |
|         NEEDBITS(8)
 | |
|         n += (unsigned)b & 0xff;
 | |
|         DUMPBITS(8)
 | |
|       }
 | |
| 
 | |
|       /* do the copy */
 | |
|       s -= n;
 | |
|       do {
 | |
| #ifdef DLL
 | |
|         if (G.redirect_data)  /* &= w/ wsize not needed and wrong if redirect */
 | |
|           n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | |
|         else
 | |
| #endif
 | |
|         n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | |
|         if (u && w <= d)
 | |
|         {
 | |
|           memzero(redirSlide + w, e);
 | |
|           w += e;
 | |
|           d += e;
 | |
|         }
 | |
|         else
 | |
| #ifndef NOMEMCPY
 | |
|           if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | |
|           {
 | |
|             memcpy(redirSlide + w, redirSlide + d, e);
 | |
|             w += e;
 | |
|             d += e;
 | |
|           }
 | |
|           else                  /* do it slow to avoid memcpy() overlap */
 | |
| #endif /* !NOMEMCPY */
 | |
|             do {
 | |
|               redirSlide[w++] = redirSlide[d++];
 | |
|             } while (--e);
 | |
|         if (w == wsize)
 | |
|         {
 | |
|           flush(__G__ redirSlide, (ulg)w, 0);
 | |
|           w = u = 0;
 | |
|         }
 | |
|       } while (n);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   /* flush out redirSlide */
 | |
|   flush(__G__ redirSlide, (ulg)w, 0);
 | |
|   if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | |
|   {                        /* sometimes read one too many:  k>>3 compensates */
 | |
|     G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | |
|     return 5;
 | |
|   }
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| static int explode_lit4(__G__ tb, tl, td, bb, bl, bd)
 | |
|      __GDEF
 | |
| struct huft *tb, *tl, *td;      /* literal, length, and distance tables */
 | |
| int bb, bl, bd;                 /* number of bits decoded by those */
 | |
| /* Decompress the imploded data using coded literals and a 4K sliding
 | |
|    window. */
 | |
| {
 | |
|   long s;               /* bytes to decompress */
 | |
|   register unsigned e;  /* table entry flag/number of extra bits */
 | |
|   unsigned n, d;        /* length and index for copy */
 | |
|   unsigned w;           /* current window position */
 | |
|   struct huft *t;       /* pointer to table entry */
 | |
|   unsigned mb, ml, md;  /* masks for bb, bl, and bd bits */
 | |
|   register ulg b;       /* bit buffer */
 | |
|   register unsigned k;  /* number of bits in bit buffer */
 | |
|   unsigned u;           /* true if unflushed */
 | |
| 
 | |
| 
 | |
|   /* explode the coded data */
 | |
|   b = k = w = 0;                /* initialize bit buffer, window */
 | |
|   u = 1;                        /* buffer unflushed */
 | |
|   mb = mask_bits[bb];           /* precompute masks for speed */
 | |
|   ml = mask_bits[bl];
 | |
|   md = mask_bits[bd];
 | |
|   s = G.ucsize;
 | |
|   while (s > 0)                 /* do until ucsize bytes uncompressed */
 | |
|   {
 | |
|     NEEDBITS(1)
 | |
|     if (b & 1)                  /* then literal--decode it */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       s--;
 | |
|       NEEDBITS((unsigned)bb)    /* get coded literal */
 | |
|       if ((e = (t = tb + ((~(unsigned)b) & mb))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       redirSlide[w++] = (uch)t->v.n;
 | |
|       if (w == wsize)
 | |
|       {
 | |
|         flush(__G__ redirSlide, (ulg)w, 0);
 | |
|         w = u = 0;
 | |
|       }
 | |
|     }
 | |
|     else                        /* else distance/length */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       NEEDBITS(6)               /* get distance low bits */
 | |
|       d = (unsigned)b & 0x3f;
 | |
|       DUMPBITS(6)
 | |
|       NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | |
|       if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       d = w - d - t->v.n;       /* construct offset */
 | |
|       NEEDBITS((unsigned)bl)    /* get coded length */
 | |
|       if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       n = t->v.n;
 | |
|       if (e)                    /* get length extra bits */
 | |
|       {
 | |
|         NEEDBITS(8)
 | |
|         n += (unsigned)b & 0xff;
 | |
|         DUMPBITS(8)
 | |
|       }
 | |
| 
 | |
|       /* do the copy */
 | |
|       s -= n;
 | |
|       do {
 | |
| #ifdef DLL
 | |
|         if (G.redirect_data)  /* &= w/ wsize not needed and wrong if redirect */
 | |
|           n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | |
|         else
 | |
| #endif
 | |
|         n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | |
|         if (u && w <= d)
 | |
|         {
 | |
|           memzero(redirSlide + w, e);
 | |
|           w += e;
 | |
|           d += e;
 | |
|         }
 | |
|         else
 | |
| #ifndef NOMEMCPY
 | |
|           if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | |
|           {
 | |
|             memcpy(redirSlide + w, redirSlide + d, e);
 | |
|             w += e;
 | |
|             d += e;
 | |
|           }
 | |
|           else                  /* do it slow to avoid memcpy() overlap */
 | |
| #endif /* !NOMEMCPY */
 | |
|             do {
 | |
|               redirSlide[w++] = redirSlide[d++];
 | |
|             } while (--e);
 | |
|         if (w == wsize)
 | |
|         {
 | |
|           flush(__G__ redirSlide, (ulg)w, 0);
 | |
|           w = u = 0;
 | |
|         }
 | |
|       } while (n);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   /* flush out redirSlide */
 | |
|   flush(__G__ redirSlide, (ulg)w, 0);
 | |
|   if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | |
|   {                        /* sometimes read one too many:  k>>3 compensates */
 | |
|     G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | |
|     return 5;
 | |
|   }
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| static int explode_nolit8(__G__ tl, td, bl, bd)
 | |
|      __GDEF
 | |
| struct huft *tl, *td;   /* length and distance decoder tables */
 | |
| int bl, bd;             /* number of bits decoded by tl[] and td[] */
 | |
| /* Decompress the imploded data using uncoded literals and an 8K sliding
 | |
|    window. */
 | |
| {
 | |
|   long s;               /* bytes to decompress */
 | |
|   register unsigned e;  /* table entry flag/number of extra bits */
 | |
|   unsigned n, d;        /* length and index for copy */
 | |
|   unsigned w;           /* current window position */
 | |
|   struct huft *t;       /* pointer to table entry */
 | |
|   unsigned ml, md;      /* masks for bl and bd bits */
 | |
|   register ulg b;       /* bit buffer */
 | |
|   register unsigned k;  /* number of bits in bit buffer */
 | |
|   unsigned u;           /* true if unflushed */
 | |
| 
 | |
| 
 | |
|   /* explode the coded data */
 | |
|   b = k = w = 0;                /* initialize bit buffer, window */
 | |
|   u = 1;                        /* buffer unflushed */
 | |
|   ml = mask_bits[bl];           /* precompute masks for speed */
 | |
|   md = mask_bits[bd];
 | |
|   s = G.ucsize;
 | |
|   while (s > 0)                 /* do until ucsize bytes uncompressed */
 | |
|   {
 | |
|     NEEDBITS(1)
 | |
|     if (b & 1)                  /* then literal--get eight bits */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       s--;
 | |
|       NEEDBITS(8)
 | |
|       redirSlide[w++] = (uch)b;
 | |
|       if (w == wsize)
 | |
|       {
 | |
|         flush(__G__ redirSlide, (ulg)w, 0);
 | |
|         w = u = 0;
 | |
|       }
 | |
|       DUMPBITS(8)
 | |
|     }
 | |
|     else                        /* else distance/length */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       NEEDBITS(7)               /* get distance low bits */
 | |
|       d = (unsigned)b & 0x7f;
 | |
|       DUMPBITS(7)
 | |
|       NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | |
|       if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       d = w - d - t->v.n;       /* construct offset */
 | |
|       NEEDBITS((unsigned)bl)    /* get coded length */
 | |
|       if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       n = t->v.n;
 | |
|       if (e)                    /* get length extra bits */
 | |
|       {
 | |
|         NEEDBITS(8)
 | |
|         n += (unsigned)b & 0xff;
 | |
|         DUMPBITS(8)
 | |
|       }
 | |
| 
 | |
|       /* do the copy */
 | |
|       s -= n;
 | |
|       do {
 | |
| #ifdef DLL
 | |
|         if (G.redirect_data)  /* &= w/ wsize not needed and wrong if redirect */
 | |
|           n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | |
|         else
 | |
| #endif
 | |
|         n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | |
|         if (u && w <= d)
 | |
|         {
 | |
|           memzero(redirSlide + w, e);
 | |
|           w += e;
 | |
|           d += e;
 | |
|         }
 | |
|         else
 | |
| #ifndef NOMEMCPY
 | |
|           if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | |
|           {
 | |
|             memcpy(redirSlide + w, redirSlide + d, e);
 | |
|             w += e;
 | |
|             d += e;
 | |
|           }
 | |
|           else                  /* do it slow to avoid memcpy() overlap */
 | |
| #endif /* !NOMEMCPY */
 | |
|             do {
 | |
|               redirSlide[w++] = redirSlide[d++];
 | |
|             } while (--e);
 | |
|         if (w == wsize)
 | |
|         {
 | |
|           flush(__G__ redirSlide, (ulg)w, 0);
 | |
|           w = u = 0;
 | |
|         }
 | |
|       } while (n);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   /* flush out redirSlide */
 | |
|   flush(__G__ redirSlide, (ulg)w, 0);
 | |
|   if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | |
|   {                        /* sometimes read one too many:  k>>3 compensates */
 | |
|     G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | |
|     return 5;
 | |
|   }
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| static int explode_nolit4(__G__ tl, td, bl, bd)
 | |
|      __GDEF
 | |
| struct huft *tl, *td;   /* length and distance decoder tables */
 | |
| int bl, bd;             /* number of bits decoded by tl[] and td[] */
 | |
| /* Decompress the imploded data using uncoded literals and a 4K sliding
 | |
|    window. */
 | |
| {
 | |
|   long s;               /* bytes to decompress */
 | |
|   register unsigned e;  /* table entry flag/number of extra bits */
 | |
|   unsigned n, d;        /* length and index for copy */
 | |
|   unsigned w;           /* current window position */
 | |
|   struct huft *t;       /* pointer to table entry */
 | |
|   unsigned ml, md;      /* masks for bl and bd bits */
 | |
|   register ulg b;       /* bit buffer */
 | |
|   register unsigned k;  /* number of bits in bit buffer */
 | |
|   unsigned u;           /* true if unflushed */
 | |
| 
 | |
| 
 | |
|   /* explode the coded data */
 | |
|   b = k = w = 0;                /* initialize bit buffer, window */
 | |
|   u = 1;                        /* buffer unflushed */
 | |
|   ml = mask_bits[bl];           /* precompute masks for speed */
 | |
|   md = mask_bits[bd];
 | |
|   s = G.ucsize;
 | |
|   while (s > 0)                 /* do until ucsize bytes uncompressed */
 | |
|   {
 | |
|     NEEDBITS(1)
 | |
|     if (b & 1)                  /* then literal--get eight bits */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       s--;
 | |
|       NEEDBITS(8)
 | |
|       redirSlide[w++] = (uch)b;
 | |
|       if (w == wsize)
 | |
|       {
 | |
|         flush(__G__ redirSlide, (ulg)w, 0);
 | |
|         w = u = 0;
 | |
|       }
 | |
|       DUMPBITS(8)
 | |
|     }
 | |
|     else                        /* else distance/length */
 | |
|     {
 | |
|       DUMPBITS(1)
 | |
|       NEEDBITS(6)               /* get distance low bits */
 | |
|       d = (unsigned)b & 0x3f;
 | |
|       DUMPBITS(6)
 | |
|       NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | |
|       if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       d = w - d - t->v.n;       /* construct offset */
 | |
|       NEEDBITS((unsigned)bl)    /* get coded length */
 | |
|       if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | |
|         do {
 | |
|           if (e == 99)
 | |
|             return 1;
 | |
|           DUMPBITS(t->b)
 | |
|           e -= 16;
 | |
|           NEEDBITS(e)
 | |
|         } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | |
|       DUMPBITS(t->b)
 | |
|       n = t->v.n;
 | |
|       if (e)                    /* get length extra bits */
 | |
|       {
 | |
|         NEEDBITS(8)
 | |
|         n += (unsigned)b & 0xff;
 | |
|         DUMPBITS(8)
 | |
|       }
 | |
| 
 | |
|       /* do the copy */
 | |
|       s -= n;
 | |
|       do {
 | |
| #ifdef DLL
 | |
|         if (G.redirect_data)  /* &= w/ wsize not needed and wrong if redirect */
 | |
|           n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | |
|         else
 | |
| #endif
 | |
|         n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | |
|         if (u && w <= d)
 | |
|         {
 | |
|           memzero(redirSlide + w, e);
 | |
|           w += e;
 | |
|           d += e;
 | |
|         }
 | |
|         else
 | |
| #ifndef NOMEMCPY
 | |
|           if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | |
|           {
 | |
|             memcpy(redirSlide + w, redirSlide + d, e);
 | |
|             w += e;
 | |
|             d += e;
 | |
|           }
 | |
|           else                  /* do it slow to avoid memcpy() overlap */
 | |
| #endif /* !NOMEMCPY */
 | |
|             do {
 | |
|               redirSlide[w++] = redirSlide[d++];
 | |
|             } while (--e);
 | |
|         if (w == wsize)
 | |
|         {
 | |
|           flush(__G__ redirSlide, (ulg)w, 0);
 | |
|           w = u = 0;
 | |
|         }
 | |
|       } while (n);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   /* flush out redirSlide */
 | |
|   flush(__G__ redirSlide, (ulg)w, 0);
 | |
|   if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | |
|   {                        /* sometimes read one too many:  k>>3 compensates */
 | |
|     G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | |
|     return 5;
 | |
|   }
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| int explode(__G)
 | |
|      __GDEF
 | |
| /* Explode an imploded compressed stream.  Based on the general purpose
 | |
|    bit flag, decide on coded or uncoded literals, and an 8K or 4K sliding
 | |
|    window.  Construct the literal (if any), length, and distance codes and
 | |
|    the tables needed to decode them (using huft_build() from inflate.c),
 | |
|    and call the appropriate routine for the type of data in the remainder
 | |
|    of the stream.  The four routines are nearly identical, differing only
 | |
|    in whether the literal is decoded or simply read in, and in how many
 | |
|    bits are read in, uncoded, for the low distance bits. */
 | |
| {
 | |
|   unsigned r;           /* return codes */
 | |
|   struct huft *tb;      /* literal code table */
 | |
|   struct huft *tl;      /* length code table */
 | |
|   struct huft *td;      /* distance code table */
 | |
|   int bb;               /* bits for tb */
 | |
|   int bl;               /* bits for tl */
 | |
|   int bd;               /* bits for td */
 | |
|   unsigned l[256];      /* bit lengths for codes */
 | |
| 
 | |
| #ifdef DLL
 | |
|   if (G.redirect_data)
 | |
|     wsize = G.redirect_size, redirSlide = G.redirect_buffer;
 | |
|   else
 | |
|     wsize = WSIZE, redirSlide = slide;
 | |
| #endif
 | |
| 
 | |
|   /* Tune base table sizes.  Note: I thought that to truly optimize speed,
 | |
|      I would have to select different bl, bd, and bb values for different
 | |
|      compressed file sizes.  I was suprised to find out the the values of
 | |
|      7, 7, and 9 worked best over a very wide range of sizes, except that
 | |
|      bd = 8 worked marginally better for large compressed sizes. */
 | |
|   bl = 7;
 | |
|   bd = (G.csize + G.incnt) > 200000L ? 8 : 7;
 | |
| 
 | |
| 
 | |
|   /* With literal tree--minimum match length is 3 */
 | |
|   G.hufts = 0;                    /* initialize huft's malloc'ed */
 | |
|   if (G.lrec.general_purpose_bit_flag & 4)
 | |
|   {
 | |
|     bb = 9;                     /* base table size for literals */
 | |
|     if ((r = get_tree(__G__ l, 256)) != 0)
 | |
|       return (int)r;
 | |
|     if ((r = huft_build(__G__ l, 256, 256, NULL, NULL, &tb, &bb)) != 0)
 | |
|     {
 | |
|       if (r == 1)
 | |
|         huft_free(tb);
 | |
|       return (int)r;
 | |
|     }
 | |
|     if ((r = get_tree(__G__ l, 64)) != 0)
 | |
|       return (int)r;
 | |
|     if ((r = huft_build(__G__ l, 64, 0, cplen3, extra, &tl, &bl)) != 0)
 | |
|     {
 | |
|       if (r == 1)
 | |
|         huft_free(tl);
 | |
|       huft_free(tb);
 | |
|       return (int)r;
 | |
|     }
 | |
|     if ((r = get_tree(__G__ l, 64)) != 0)
 | |
|       return (int)r;
 | |
|     if (G.lrec.general_purpose_bit_flag & 2)      /* true if 8K */
 | |
|     {
 | |
|       if ((r = huft_build(__G__ l, 64, 0, cpdist8, extra, &td, &bd)) != 0)
 | |
|       {
 | |
|         if (r == 1)
 | |
|           huft_free(td);
 | |
|         huft_free(tl);
 | |
|         huft_free(tb);
 | |
|         return (int)r;
 | |
|       }
 | |
|       r = explode_lit8(__G__ tb, tl, td, bb, bl, bd);
 | |
|     }
 | |
|     else                                        /* else 4K */
 | |
|     {
 | |
|       if ((r = huft_build(__G__ l, 64, 0, cpdist4, extra, &td, &bd)) != 0)
 | |
|       {
 | |
|         if (r == 1)
 | |
|           huft_free(td);
 | |
|         huft_free(tl);
 | |
|         huft_free(tb);
 | |
|         return (int)r;
 | |
|       }
 | |
|       r = explode_lit4(__G__ tb, tl, td, bb, bl, bd);
 | |
|     }
 | |
|     huft_free(td);
 | |
|     huft_free(tl);
 | |
|     huft_free(tb);
 | |
|   }
 | |
|   else
 | |
| 
 | |
| 
 | |
|   /* No literal tree--minimum match length is 2 */
 | |
|   {
 | |
|     if ((r = get_tree(__G__ l, 64)) != 0)
 | |
|       return (int)r;
 | |
|     if ((r = huft_build(__G__ l, 64, 0, cplen2, extra, &tl, &bl)) != 0)
 | |
|     {
 | |
|       if (r == 1)
 | |
|         huft_free(tl);
 | |
|       return (int)r;
 | |
|     }
 | |
|     if ((r = get_tree(__G__ l, 64)) != 0)
 | |
|       return (int)r;
 | |
|     if (G.lrec.general_purpose_bit_flag & 2)      /* true if 8K */
 | |
|     {
 | |
|       if ((r = huft_build(__G__ l, 64, 0, cpdist8, extra, &td, &bd)) != 0)
 | |
|       {
 | |
|         if (r == 1)
 | |
|           huft_free(td);
 | |
|         huft_free(tl);
 | |
|         return (int)r;
 | |
|       }
 | |
|       r = explode_nolit8(__G__ tl, td, bl, bd);
 | |
|     }
 | |
|     else                                        /* else 4K */
 | |
|     {
 | |
|       if ((r = huft_build(__G__ l, 64, 0, cpdist4, extra, &td, &bd)) != 0)
 | |
|       {
 | |
|         if (r == 1)
 | |
|           huft_free(td);
 | |
|         huft_free(tl);
 | |
|         return (int)r;
 | |
|       }
 | |
|       r = explode_nolit4(__G__ tl, td, bl, bd);
 | |
|     }
 | |
|     huft_free(td);
 | |
|     huft_free(tl);
 | |
|   }
 | |
| #ifdef DEBUG
 | |
|   fprintf(stderr, "<%u > ", G.hufts);
 | |
| #endif /* DEBUG */
 | |
|   return (int)r;
 | |
| }
 | |
| 
 | |
| /* so explode.c and inflate.c can be compiled together into one object: */
 | |
| #undef NEXTBYTE
 | |
| #undef NEEDBITS
 | |
| #undef DUMPBITS
 |