git-svn-id: https://svn.wxwidgets.org/svn/wx/wxWidgets/trunk@7748 c3d73ce0-8a6f-49c7-b76d-6d57e0e08775
		
			
				
	
	
		
			872 lines
		
	
	
		
			30 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			872 lines
		
	
	
		
			30 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/* explode.c -- put in the public domain by Mark Adler
 | 
						|
   version c15, 6 July 1996 */
 | 
						|
 | 
						|
 | 
						|
/* You can do whatever you like with this source file, though I would
 | 
						|
   prefer that if you modify it and redistribute it that you include
 | 
						|
   comments to that effect with your name and the date.  Thank you.
 | 
						|
 | 
						|
   History:
 | 
						|
   vers    date          who           what
 | 
						|
   ----  ---------  --------------  ------------------------------------
 | 
						|
    c1   30 Mar 92  M. Adler        explode that uses huft_build from inflate
 | 
						|
                                    (this gives over a 70% speed improvement
 | 
						|
                                    over the original unimplode.c, which
 | 
						|
                                    decoded a bit at a time)
 | 
						|
    c2    4 Apr 92  M. Adler        fixed bug for file sizes a multiple of 32k.
 | 
						|
    c3   10 Apr 92  M. Adler        added a little memory tracking if DEBUG
 | 
						|
    c4   11 Apr 92  M. Adler        added NOMEMCPY do kill use of memcpy()
 | 
						|
    c5   21 Apr 92  M. Adler        added the WSIZE #define to allow reducing
 | 
						|
                                    the 32K window size for specialized
 | 
						|
                                    applications.
 | 
						|
    c6   31 May 92  M. Adler        added typecasts to eliminate some warnings
 | 
						|
    c7   27 Jun 92  G. Roelofs      added more typecasts.
 | 
						|
    c8   17 Oct 92  G. Roelofs      changed ULONG/UWORD/byte to ulg/ush/uch.
 | 
						|
    c9   19 Jul 93  J. Bush         added more typecasts (to return values);
 | 
						|
                                    made l[256] array static for Amiga.
 | 
						|
    c10   8 Oct 93  G. Roelofs      added used_csize for diagnostics; added
 | 
						|
                                    buf and unshrink arguments to flush();
 | 
						|
                                    undef'd various macros at end for Turbo C;
 | 
						|
                                    removed NEXTBYTE macro (now in unzip.h)
 | 
						|
                                    and bytebuf variable (not used); changed
 | 
						|
                                    memset() to memzero().
 | 
						|
    c11   9 Jan 94  M. Adler        fixed incorrect used_csize calculation.
 | 
						|
    c12   9 Apr 94  G. Roelofs      fixed split comments on preprocessor lines
 | 
						|
                                    to avoid bug in Encore compiler.
 | 
						|
    c13  25 Aug 94  M. Adler        fixed distance-length comment (orig c9 fix)
 | 
						|
    c14  22 Nov 95  S. Maxwell      removed unnecessary "static" on auto array
 | 
						|
    c15   6 Jul 96  W. Haidinger    added ulg typecasts to flush() calls.
 | 
						|
    c16   8 Feb 98  C. Spieler      added ZCONST modifiers to const tables
 | 
						|
                                    and #ifdef DEBUG around debugging code.
 | 
						|
    c16b 25 Mar 98  C. Spieler      modified DLL code for slide redirection.
 | 
						|
 */
 | 
						|
 | 
						|
 | 
						|
/*
 | 
						|
   Explode imploded (PKZIP method 6 compressed) data.  This compression
 | 
						|
   method searches for as much of the current string of bytes (up to a length
 | 
						|
   of ~320) in the previous 4K or 8K bytes.  If it doesn't find any matches
 | 
						|
   (of at least length 2 or 3), it codes the next byte.  Otherwise, it codes
 | 
						|
   the length of the matched string and its distance backwards from the
 | 
						|
   current position.  Single bytes ("literals") are preceded by a one (a
 | 
						|
   single bit) and are either uncoded (the eight bits go directly into the
 | 
						|
   compressed stream for a total of nine bits) or Huffman coded with a
 | 
						|
   supplied literal code tree.  If literals are coded, then the minimum match
 | 
						|
   length is three, otherwise it is two.
 | 
						|
 | 
						|
   There are therefore four kinds of imploded streams: 8K search with coded
 | 
						|
   literals (min match = 3), 4K search with coded literals (min match = 3),
 | 
						|
   8K with uncoded literals (min match = 2), and 4K with uncoded literals
 | 
						|
   (min match = 2).  The kind of stream is identified in two bits of a
 | 
						|
   general purpose bit flag that is outside of the compressed stream.
 | 
						|
 | 
						|
   Distance-length pairs for matched strings are preceded by a zero bit (to
 | 
						|
   distinguish them from literals) and are always coded.  The distance comes
 | 
						|
   first and is either the low six (4K) or low seven (8K) bits of the
 | 
						|
   distance (uncoded), followed by the high six bits of the distance coded.
 | 
						|
   Then the length is six bits coded (0..63 + min match length), and if the
 | 
						|
   maximum such length is coded, then it's followed by another eight bits
 | 
						|
   (uncoded) to be added to the coded length.  This gives a match length
 | 
						|
   range of 2..320 or 3..321 bytes.
 | 
						|
 | 
						|
   The literal, length, and distance codes are all represented in a slightly
 | 
						|
   compressed form themselves.  What is sent are the lengths of the codes for
 | 
						|
   each value, which is sufficient to construct the codes.  Each byte of the
 | 
						|
   code representation is the code length (the low four bits representing
 | 
						|
   1..16), and the number of values sequentially with that length (the high
 | 
						|
   four bits also representing 1..16).  There are 256 literal code values (if
 | 
						|
   literals are coded), 64 length code values, and 64 distance code values,
 | 
						|
   in that order at the beginning of the compressed stream.  Each set of code
 | 
						|
   values is preceded (redundantly) with a byte indicating how many bytes are
 | 
						|
   in the code description that follows, in the range 1..256.
 | 
						|
 | 
						|
   The codes themselves are decoded using tables made by huft_build() from
 | 
						|
   the bit lengths.  That routine and its comments are in the inflate.c
 | 
						|
   module.
 | 
						|
 */
 | 
						|
 | 
						|
#define UNZIP_INTERNAL
 | 
						|
#include "unzip.h"      /* must supply slide[] (uch) array and NEXTBYTE macro */
 | 
						|
 | 
						|
#ifndef WSIZE
 | 
						|
#  define WSIZE 0x8000  /* window size--must be a power of two, and */
 | 
						|
#endif                  /* at least 8K for zip's implode method */
 | 
						|
 | 
						|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
 | 
						|
#  define wsize G._wsize
 | 
						|
#else
 | 
						|
#  define wsize WSIZE
 | 
						|
#endif
 | 
						|
 | 
						|
/* routines here */
 | 
						|
static int get_tree OF((__GPRO__ unsigned *l, unsigned n));
 | 
						|
static int explode_lit8 OF((__GPRO__ struct huft *tb, struct huft *tl,
 | 
						|
                            struct huft *td, int bb, int bl, int bd));
 | 
						|
static int explode_lit4 OF((__GPRO__ struct huft *tb, struct huft *tl,
 | 
						|
                            struct huft *td, int bb, int bl, int bd));
 | 
						|
static int explode_nolit8 OF((__GPRO__ struct huft *tl, struct huft *td,
 | 
						|
                              int bl, int bd));
 | 
						|
static int explode_nolit4 OF((__GPRO__ struct huft *tl, struct huft *td,
 | 
						|
                              int bl, int bd));
 | 
						|
int explode OF((__GPRO));
 | 
						|
 | 
						|
 | 
						|
/* The implode algorithm uses a sliding 4K or 8K byte window on the
 | 
						|
   uncompressed stream to find repeated byte strings.  This is implemented
 | 
						|
   here as a circular buffer.  The index is updated simply by incrementing
 | 
						|
   and then and'ing with 0x0fff (4K-1) or 0x1fff (8K-1).  Here, the 32K
 | 
						|
   buffer of inflate is used, and it works just as well to always have
 | 
						|
   a 32K circular buffer, so the index is anded with 0x7fff.  This is
 | 
						|
   done to allow the window to also be used as the output buffer. */
 | 
						|
/* This must be supplied in an external module useable like "uch slide[8192];"
 | 
						|
   or "uch *slide;", where the latter would be malloc'ed.  In unzip, slide[]
 | 
						|
   is actually a 32K area for use by inflate, which uses a 32K sliding window.
 | 
						|
 */
 | 
						|
 | 
						|
 | 
						|
/* Tables for length and distance */
 | 
						|
static ZCONST ush cplen2[] =
 | 
						|
        {2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17,
 | 
						|
        18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34,
 | 
						|
        35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51,
 | 
						|
        52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65};
 | 
						|
static ZCONST ush cplen3[] =
 | 
						|
        {3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18,
 | 
						|
        19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35,
 | 
						|
        36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52,
 | 
						|
        53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66};
 | 
						|
static ZCONST ush extra[] =
 | 
						|
        {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 | 
						|
        0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 | 
						|
        0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
 | 
						|
        8};
 | 
						|
static ZCONST ush cpdist4[] =
 | 
						|
        {1, 65, 129, 193, 257, 321, 385, 449, 513, 577, 641, 705,
 | 
						|
        769, 833, 897, 961, 1025, 1089, 1153, 1217, 1281, 1345, 1409, 1473,
 | 
						|
        1537, 1601, 1665, 1729, 1793, 1857, 1921, 1985, 2049, 2113, 2177,
 | 
						|
        2241, 2305, 2369, 2433, 2497, 2561, 2625, 2689, 2753, 2817, 2881,
 | 
						|
        2945, 3009, 3073, 3137, 3201, 3265, 3329, 3393, 3457, 3521, 3585,
 | 
						|
        3649, 3713, 3777, 3841, 3905, 3969, 4033};
 | 
						|
static ZCONST ush cpdist8[] =
 | 
						|
        {1, 129, 257, 385, 513, 641, 769, 897, 1025, 1153, 1281,
 | 
						|
        1409, 1537, 1665, 1793, 1921, 2049, 2177, 2305, 2433, 2561, 2689,
 | 
						|
        2817, 2945, 3073, 3201, 3329, 3457, 3585, 3713, 3841, 3969, 4097,
 | 
						|
        4225, 4353, 4481, 4609, 4737, 4865, 4993, 5121, 5249, 5377, 5505,
 | 
						|
        5633, 5761, 5889, 6017, 6145, 6273, 6401, 6529, 6657, 6785, 6913,
 | 
						|
        7041, 7169, 7297, 7425, 7553, 7681, 7809, 7937, 8065};
 | 
						|
 | 
						|
 | 
						|
/* Macros for inflate() bit peeking and grabbing.
 | 
						|
   The usage is:
 | 
						|
 | 
						|
        NEEDBITS(j)
 | 
						|
        x = b & mask_bits[j];
 | 
						|
        DUMPBITS(j)
 | 
						|
 | 
						|
   where NEEDBITS makes sure that b has at least j bits in it, and
 | 
						|
   DUMPBITS removes the bits from b.  The macros use the variable k
 | 
						|
   for the number of bits in b.  Normally, b and k are register
 | 
						|
   variables for speed.
 | 
						|
 */
 | 
						|
 | 
						|
#define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE)<<k;k+=8;}}
 | 
						|
#define DUMPBITS(n) {b>>=(n);k-=(n);}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
static int get_tree(__G__ l, n)
 | 
						|
     __GDEF
 | 
						|
unsigned *l;            /* bit lengths */
 | 
						|
unsigned n;             /* number expected */
 | 
						|
/* Get the bit lengths for a code representation from the compressed
 | 
						|
   stream.  If get_tree() returns 4, then there is an error in the data.
 | 
						|
   Otherwise zero is returned. */
 | 
						|
{
 | 
						|
  unsigned i;           /* bytes remaining in list */
 | 
						|
  unsigned k;           /* lengths entered */
 | 
						|
  unsigned j;           /* number of codes */
 | 
						|
  unsigned b;           /* bit length for those codes */
 | 
						|
 | 
						|
 | 
						|
  /* get bit lengths */
 | 
						|
  i = NEXTBYTE + 1;                     /* length/count pairs to read */
 | 
						|
  k = 0;                                /* next code */
 | 
						|
  do {
 | 
						|
    b = ((j = NEXTBYTE) & 0xf) + 1;     /* bits in code (1..16) */
 | 
						|
    j = ((j & 0xf0) >> 4) + 1;          /* codes with those bits (1..16) */
 | 
						|
    if (k + j > n)
 | 
						|
      return 4;                         /* don't overflow l[] */
 | 
						|
    do {
 | 
						|
      l[k++] = b;
 | 
						|
    } while (--j);
 | 
						|
  } while (--i);
 | 
						|
  return k != n ? 4 : 0;                /* should have read n of them */
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
static int explode_lit8(__G__ tb, tl, td, bb, bl, bd)
 | 
						|
     __GDEF
 | 
						|
struct huft *tb, *tl, *td;      /* literal, length, and distance tables */
 | 
						|
int bb, bl, bd;                 /* number of bits decoded by those */
 | 
						|
/* Decompress the imploded data using coded literals and an 8K sliding
 | 
						|
   window. */
 | 
						|
{
 | 
						|
  long s;               /* bytes to decompress */
 | 
						|
  register unsigned e;  /* table entry flag/number of extra bits */
 | 
						|
  unsigned n, d;        /* length and index for copy */
 | 
						|
  unsigned w;           /* current window position */
 | 
						|
  struct huft *t;       /* pointer to table entry */
 | 
						|
  unsigned mb, ml, md;  /* masks for bb, bl, and bd bits */
 | 
						|
  register ulg b;       /* bit buffer */
 | 
						|
  register unsigned k;  /* number of bits in bit buffer */
 | 
						|
  unsigned u;           /* true if unflushed */
 | 
						|
 | 
						|
 | 
						|
  /* explode the coded data */
 | 
						|
  b = k = w = 0;                /* initialize bit buffer, window */
 | 
						|
  u = 1;                        /* buffer unflushed */
 | 
						|
  mb = mask_bits[bb];           /* precompute masks for speed */
 | 
						|
  ml = mask_bits[bl];
 | 
						|
  md = mask_bits[bd];
 | 
						|
  s = G.ucsize;
 | 
						|
  while (s > 0)                 /* do until ucsize bytes uncompressed */
 | 
						|
  {
 | 
						|
    NEEDBITS(1)
 | 
						|
    if (b & 1)                  /* then literal--decode it */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      s--;
 | 
						|
      NEEDBITS((unsigned)bb)    /* get coded literal */
 | 
						|
      if ((e = (t = tb + ((~(unsigned)b) & mb))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      redirSlide[w++] = (uch)t->v.n;
 | 
						|
      if (w == wsize)
 | 
						|
      {
 | 
						|
        flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
        w = u = 0;
 | 
						|
      }
 | 
						|
    }
 | 
						|
    else                        /* else distance/length */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      NEEDBITS(7)               /* get distance low bits */
 | 
						|
      d = (unsigned)b & 0x7f;
 | 
						|
      DUMPBITS(7)
 | 
						|
      NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | 
						|
      if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      d = w - d - t->v.n;       /* construct offset */
 | 
						|
      NEEDBITS((unsigned)bl)    /* get coded length */
 | 
						|
      if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      n = t->v.n;
 | 
						|
      if (e)                    /* get length extra bits */
 | 
						|
      {
 | 
						|
        NEEDBITS(8)
 | 
						|
        n += (unsigned)b & 0xff;
 | 
						|
        DUMPBITS(8)
 | 
						|
      }
 | 
						|
 | 
						|
      /* do the copy */
 | 
						|
      s -= n;
 | 
						|
      do {
 | 
						|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
 | 
						|
        if (G.redirect_slide) {
 | 
						|
          /* &= w/ wsize not needed and wrong if redirect */
 | 
						|
          if (d >= wsize)
 | 
						|
            return 1;
 | 
						|
          n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | 
						|
        } else
 | 
						|
#endif
 | 
						|
        n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | 
						|
        if (u && w <= d)
 | 
						|
        {
 | 
						|
          memzero(redirSlide + w, e);
 | 
						|
          w += e;
 | 
						|
          d += e;
 | 
						|
        }
 | 
						|
        else
 | 
						|
#ifndef NOMEMCPY
 | 
						|
          if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | 
						|
          {
 | 
						|
            memcpy(redirSlide + w, redirSlide + d, e);
 | 
						|
            w += e;
 | 
						|
            d += e;
 | 
						|
          }
 | 
						|
          else                  /* do it slow to avoid memcpy() overlap */
 | 
						|
#endif /* !NOMEMCPY */
 | 
						|
            do {
 | 
						|
              redirSlide[w++] = redirSlide[d++];
 | 
						|
            } while (--e);
 | 
						|
        if (w == wsize)
 | 
						|
        {
 | 
						|
          flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
          w = u = 0;
 | 
						|
        }
 | 
						|
      } while (n);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /* flush out redirSlide */
 | 
						|
  flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
  if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | 
						|
  {                        /* sometimes read one too many:  k>>3 compensates */
 | 
						|
    G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | 
						|
    return 5;
 | 
						|
  }
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
static int explode_lit4(__G__ tb, tl, td, bb, bl, bd)
 | 
						|
     __GDEF
 | 
						|
struct huft *tb, *tl, *td;      /* literal, length, and distance tables */
 | 
						|
int bb, bl, bd;                 /* number of bits decoded by those */
 | 
						|
/* Decompress the imploded data using coded literals and a 4K sliding
 | 
						|
   window. */
 | 
						|
{
 | 
						|
  long s;               /* bytes to decompress */
 | 
						|
  register unsigned e;  /* table entry flag/number of extra bits */
 | 
						|
  unsigned n, d;        /* length and index for copy */
 | 
						|
  unsigned w;           /* current window position */
 | 
						|
  struct huft *t;       /* pointer to table entry */
 | 
						|
  unsigned mb, ml, md;  /* masks for bb, bl, and bd bits */
 | 
						|
  register ulg b;       /* bit buffer */
 | 
						|
  register unsigned k;  /* number of bits in bit buffer */
 | 
						|
  unsigned u;           /* true if unflushed */
 | 
						|
 | 
						|
 | 
						|
  /* explode the coded data */
 | 
						|
  b = k = w = 0;                /* initialize bit buffer, window */
 | 
						|
  u = 1;                        /* buffer unflushed */
 | 
						|
  mb = mask_bits[bb];           /* precompute masks for speed */
 | 
						|
  ml = mask_bits[bl];
 | 
						|
  md = mask_bits[bd];
 | 
						|
  s = G.ucsize;
 | 
						|
  while (s > 0)                 /* do until ucsize bytes uncompressed */
 | 
						|
  {
 | 
						|
    NEEDBITS(1)
 | 
						|
    if (b & 1)                  /* then literal--decode it */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      s--;
 | 
						|
      NEEDBITS((unsigned)bb)    /* get coded literal */
 | 
						|
      if ((e = (t = tb + ((~(unsigned)b) & mb))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      redirSlide[w++] = (uch)t->v.n;
 | 
						|
      if (w == wsize)
 | 
						|
      {
 | 
						|
        flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
        w = u = 0;
 | 
						|
      }
 | 
						|
    }
 | 
						|
    else                        /* else distance/length */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      NEEDBITS(6)               /* get distance low bits */
 | 
						|
      d = (unsigned)b & 0x3f;
 | 
						|
      DUMPBITS(6)
 | 
						|
      NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | 
						|
      if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      d = w - d - t->v.n;       /* construct offset */
 | 
						|
      NEEDBITS((unsigned)bl)    /* get coded length */
 | 
						|
      if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      n = t->v.n;
 | 
						|
      if (e)                    /* get length extra bits */
 | 
						|
      {
 | 
						|
        NEEDBITS(8)
 | 
						|
        n += (unsigned)b & 0xff;
 | 
						|
        DUMPBITS(8)
 | 
						|
      }
 | 
						|
 | 
						|
      /* do the copy */
 | 
						|
      s -= n;
 | 
						|
      do {
 | 
						|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
 | 
						|
        if (G.redirect_slide) {
 | 
						|
          /* &= w/ wsize not needed and wrong if redirect */
 | 
						|
          if (d >= wsize)
 | 
						|
            return 1;
 | 
						|
          n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | 
						|
        } else
 | 
						|
#endif
 | 
						|
        n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | 
						|
        if (u && w <= d)
 | 
						|
        {
 | 
						|
          memzero(redirSlide + w, e);
 | 
						|
          w += e;
 | 
						|
          d += e;
 | 
						|
        }
 | 
						|
        else
 | 
						|
#ifndef NOMEMCPY
 | 
						|
          if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | 
						|
          {
 | 
						|
            memcpy(redirSlide + w, redirSlide + d, e);
 | 
						|
            w += e;
 | 
						|
            d += e;
 | 
						|
          }
 | 
						|
          else                  /* do it slow to avoid memcpy() overlap */
 | 
						|
#endif /* !NOMEMCPY */
 | 
						|
            do {
 | 
						|
              redirSlide[w++] = redirSlide[d++];
 | 
						|
            } while (--e);
 | 
						|
        if (w == wsize)
 | 
						|
        {
 | 
						|
          flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
          w = u = 0;
 | 
						|
        }
 | 
						|
      } while (n);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /* flush out redirSlide */
 | 
						|
  flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
  if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | 
						|
  {                        /* sometimes read one too many:  k>>3 compensates */
 | 
						|
    G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | 
						|
    return 5;
 | 
						|
  }
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
static int explode_nolit8(__G__ tl, td, bl, bd)
 | 
						|
     __GDEF
 | 
						|
struct huft *tl, *td;   /* length and distance decoder tables */
 | 
						|
int bl, bd;             /* number of bits decoded by tl[] and td[] */
 | 
						|
/* Decompress the imploded data using uncoded literals and an 8K sliding
 | 
						|
   window. */
 | 
						|
{
 | 
						|
  long s;               /* bytes to decompress */
 | 
						|
  register unsigned e;  /* table entry flag/number of extra bits */
 | 
						|
  unsigned n, d;        /* length and index for copy */
 | 
						|
  unsigned w;           /* current window position */
 | 
						|
  struct huft *t;       /* pointer to table entry */
 | 
						|
  unsigned ml, md;      /* masks for bl and bd bits */
 | 
						|
  register ulg b;       /* bit buffer */
 | 
						|
  register unsigned k;  /* number of bits in bit buffer */
 | 
						|
  unsigned u;           /* true if unflushed */
 | 
						|
 | 
						|
 | 
						|
  /* explode the coded data */
 | 
						|
  b = k = w = 0;                /* initialize bit buffer, window */
 | 
						|
  u = 1;                        /* buffer unflushed */
 | 
						|
  ml = mask_bits[bl];           /* precompute masks for speed */
 | 
						|
  md = mask_bits[bd];
 | 
						|
  s = G.ucsize;
 | 
						|
  while (s > 0)                 /* do until ucsize bytes uncompressed */
 | 
						|
  {
 | 
						|
    NEEDBITS(1)
 | 
						|
    if (b & 1)                  /* then literal--get eight bits */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      s--;
 | 
						|
      NEEDBITS(8)
 | 
						|
      redirSlide[w++] = (uch)b;
 | 
						|
      if (w == wsize)
 | 
						|
      {
 | 
						|
        flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
        w = u = 0;
 | 
						|
      }
 | 
						|
      DUMPBITS(8)
 | 
						|
    }
 | 
						|
    else                        /* else distance/length */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      NEEDBITS(7)               /* get distance low bits */
 | 
						|
      d = (unsigned)b & 0x7f;
 | 
						|
      DUMPBITS(7)
 | 
						|
      NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | 
						|
      if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      d = w - d - t->v.n;       /* construct offset */
 | 
						|
      NEEDBITS((unsigned)bl)    /* get coded length */
 | 
						|
      if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      n = t->v.n;
 | 
						|
      if (e)                    /* get length extra bits */
 | 
						|
      {
 | 
						|
        NEEDBITS(8)
 | 
						|
        n += (unsigned)b & 0xff;
 | 
						|
        DUMPBITS(8)
 | 
						|
      }
 | 
						|
 | 
						|
      /* do the copy */
 | 
						|
      s -= n;
 | 
						|
      do {
 | 
						|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
 | 
						|
        if (G.redirect_slide) {
 | 
						|
          /* &= w/ wsize not needed and wrong if redirect */
 | 
						|
          if (d >= wsize)
 | 
						|
            return 1;
 | 
						|
          n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | 
						|
        } else
 | 
						|
#endif
 | 
						|
        n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | 
						|
        if (u && w <= d)
 | 
						|
        {
 | 
						|
          memzero(redirSlide + w, e);
 | 
						|
          w += e;
 | 
						|
          d += e;
 | 
						|
        }
 | 
						|
        else
 | 
						|
#ifndef NOMEMCPY
 | 
						|
          if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | 
						|
          {
 | 
						|
            memcpy(redirSlide + w, redirSlide + d, e);
 | 
						|
            w += e;
 | 
						|
            d += e;
 | 
						|
          }
 | 
						|
          else                  /* do it slow to avoid memcpy() overlap */
 | 
						|
#endif /* !NOMEMCPY */
 | 
						|
            do {
 | 
						|
              redirSlide[w++] = redirSlide[d++];
 | 
						|
            } while (--e);
 | 
						|
        if (w == wsize)
 | 
						|
        {
 | 
						|
          flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
          w = u = 0;
 | 
						|
        }
 | 
						|
      } while (n);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /* flush out redirSlide */
 | 
						|
  flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
  if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | 
						|
  {                        /* sometimes read one too many:  k>>3 compensates */
 | 
						|
    G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | 
						|
    return 5;
 | 
						|
  }
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
static int explode_nolit4(__G__ tl, td, bl, bd)
 | 
						|
     __GDEF
 | 
						|
struct huft *tl, *td;   /* length and distance decoder tables */
 | 
						|
int bl, bd;             /* number of bits decoded by tl[] and td[] */
 | 
						|
/* Decompress the imploded data using uncoded literals and a 4K sliding
 | 
						|
   window. */
 | 
						|
{
 | 
						|
  long s;               /* bytes to decompress */
 | 
						|
  register unsigned e;  /* table entry flag/number of extra bits */
 | 
						|
  unsigned n, d;        /* length and index for copy */
 | 
						|
  unsigned w;           /* current window position */
 | 
						|
  struct huft *t;       /* pointer to table entry */
 | 
						|
  unsigned ml, md;      /* masks for bl and bd bits */
 | 
						|
  register ulg b;       /* bit buffer */
 | 
						|
  register unsigned k;  /* number of bits in bit buffer */
 | 
						|
  unsigned u;           /* true if unflushed */
 | 
						|
 | 
						|
 | 
						|
  /* explode the coded data */
 | 
						|
  b = k = w = 0;                /* initialize bit buffer, window */
 | 
						|
  u = 1;                        /* buffer unflushed */
 | 
						|
  ml = mask_bits[bl];           /* precompute masks for speed */
 | 
						|
  md = mask_bits[bd];
 | 
						|
  s = G.ucsize;
 | 
						|
  while (s > 0)                 /* do until ucsize bytes uncompressed */
 | 
						|
  {
 | 
						|
    NEEDBITS(1)
 | 
						|
    if (b & 1)                  /* then literal--get eight bits */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      s--;
 | 
						|
      NEEDBITS(8)
 | 
						|
      redirSlide[w++] = (uch)b;
 | 
						|
      if (w == wsize)
 | 
						|
      {
 | 
						|
        flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
        w = u = 0;
 | 
						|
      }
 | 
						|
      DUMPBITS(8)
 | 
						|
    }
 | 
						|
    else                        /* else distance/length */
 | 
						|
    {
 | 
						|
      DUMPBITS(1)
 | 
						|
      NEEDBITS(6)               /* get distance low bits */
 | 
						|
      d = (unsigned)b & 0x3f;
 | 
						|
      DUMPBITS(6)
 | 
						|
      NEEDBITS((unsigned)bd)    /* get coded distance high bits */
 | 
						|
      if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      d = w - d - t->v.n;       /* construct offset */
 | 
						|
      NEEDBITS((unsigned)bl)    /* get coded length */
 | 
						|
      if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
 | 
						|
        do {
 | 
						|
          if (e == 99)
 | 
						|
            return 1;
 | 
						|
          DUMPBITS(t->b)
 | 
						|
          e -= 16;
 | 
						|
          NEEDBITS(e)
 | 
						|
        } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
 | 
						|
      DUMPBITS(t->b)
 | 
						|
      n = t->v.n;
 | 
						|
      if (e)                    /* get length extra bits */
 | 
						|
      {
 | 
						|
        NEEDBITS(8)
 | 
						|
        n += (unsigned)b & 0xff;
 | 
						|
        DUMPBITS(8)
 | 
						|
      }
 | 
						|
 | 
						|
      /* do the copy */
 | 
						|
      s -= n;
 | 
						|
      do {
 | 
						|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
 | 
						|
        if (G.redirect_slide) {
 | 
						|
          /* &= w/ wsize not needed and wrong if redirect */
 | 
						|
          if (d >= wsize)
 | 
						|
            return 1;
 | 
						|
          n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
 | 
						|
        } else
 | 
						|
#endif
 | 
						|
        n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
 | 
						|
        if (u && w <= d)
 | 
						|
        {
 | 
						|
          memzero(redirSlide + w, e);
 | 
						|
          w += e;
 | 
						|
          d += e;
 | 
						|
        }
 | 
						|
        else
 | 
						|
#ifndef NOMEMCPY
 | 
						|
          if (w - d >= e)       /* (this test assumes unsigned comparison) */
 | 
						|
          {
 | 
						|
            memcpy(redirSlide + w, redirSlide + d, e);
 | 
						|
            w += e;
 | 
						|
            d += e;
 | 
						|
          }
 | 
						|
          else                  /* do it slow to avoid memcpy() overlap */
 | 
						|
#endif /* !NOMEMCPY */
 | 
						|
            do {
 | 
						|
              redirSlide[w++] = redirSlide[d++];
 | 
						|
            } while (--e);
 | 
						|
        if (w == wsize)
 | 
						|
        {
 | 
						|
          flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
          w = u = 0;
 | 
						|
        }
 | 
						|
      } while (n);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /* flush out redirSlide */
 | 
						|
  flush(__G__ redirSlide, (ulg)w, 0);
 | 
						|
  if (G.csize + G.incnt + (k >> 3))   /* should have read csize bytes, but */
 | 
						|
  {                        /* sometimes read one too many:  k>>3 compensates */
 | 
						|
    G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
 | 
						|
    return 5;
 | 
						|
  }
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
int explode(__G)
 | 
						|
     __GDEF
 | 
						|
/* Explode an imploded compressed stream.  Based on the general purpose
 | 
						|
   bit flag, decide on coded or uncoded literals, and an 8K or 4K sliding
 | 
						|
   window.  Construct the literal (if any), length, and distance codes and
 | 
						|
   the tables needed to decode them (using huft_build() from inflate.c),
 | 
						|
   and call the appropriate routine for the type of data in the remainder
 | 
						|
   of the stream.  The four routines are nearly identical, differing only
 | 
						|
   in whether the literal is decoded or simply read in, and in how many
 | 
						|
   bits are read in, uncoded, for the low distance bits. */
 | 
						|
{
 | 
						|
  unsigned r;           /* return codes */
 | 
						|
  struct huft *tb;      /* literal code table */
 | 
						|
  struct huft *tl;      /* length code table */
 | 
						|
  struct huft *td;      /* distance code table */
 | 
						|
  int bb;               /* bits for tb */
 | 
						|
  int bl;               /* bits for tl */
 | 
						|
  int bd;               /* bits for td */
 | 
						|
  unsigned l[256];      /* bit lengths for codes */
 | 
						|
 | 
						|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
 | 
						|
  if (G.redirect_slide)
 | 
						|
    wsize = G.redirect_size, redirSlide = G.redirect_buffer;
 | 
						|
  else
 | 
						|
    wsize = WSIZE, redirSlide = slide;
 | 
						|
#endif
 | 
						|
 | 
						|
  /* Tune base table sizes.  Note: I thought that to truly optimize speed,
 | 
						|
     I would have to select different bl, bd, and bb values for different
 | 
						|
     compressed file sizes.  I was surprised to find out that the values of
 | 
						|
     7, 7, and 9 worked best over a very wide range of sizes, except that
 | 
						|
     bd = 8 worked marginally better for large compressed sizes. */
 | 
						|
  bl = 7;
 | 
						|
  bd = (G.csize + G.incnt) > 200000L ? 8 : 7;
 | 
						|
 | 
						|
 | 
						|
  /* With literal tree--minimum match length is 3 */
 | 
						|
#ifdef DEBUG
 | 
						|
  G.hufts = 0;                    /* initialize huft's malloc'ed */
 | 
						|
#endif
 | 
						|
  if (G.lrec.general_purpose_bit_flag & 4)
 | 
						|
  {
 | 
						|
    bb = 9;                     /* base table size for literals */
 | 
						|
    if ((r = get_tree(__G__ l, 256)) != 0)
 | 
						|
      return (int)r;
 | 
						|
    if ((r = huft_build(__G__ l, 256, 256, NULL, NULL, &tb, &bb)) != 0)
 | 
						|
    {
 | 
						|
      if (r == 1)
 | 
						|
        huft_free(tb);
 | 
						|
      return (int)r;
 | 
						|
    }
 | 
						|
    if ((r = get_tree(__G__ l, 64)) != 0)
 | 
						|
      return (int)r;
 | 
						|
    if ((r = huft_build(__G__ l, 64, 0, cplen3, extra, &tl, &bl)) != 0)
 | 
						|
    {
 | 
						|
      if (r == 1)
 | 
						|
        huft_free(tl);
 | 
						|
      huft_free(tb);
 | 
						|
      return (int)r;
 | 
						|
    }
 | 
						|
    if ((r = get_tree(__G__ l, 64)) != 0)
 | 
						|
      return (int)r;
 | 
						|
    if (G.lrec.general_purpose_bit_flag & 2)      /* true if 8K */
 | 
						|
    {
 | 
						|
      if ((r = huft_build(__G__ l, 64, 0, cpdist8, extra, &td, &bd)) != 0)
 | 
						|
      {
 | 
						|
        if (r == 1)
 | 
						|
          huft_free(td);
 | 
						|
        huft_free(tl);
 | 
						|
        huft_free(tb);
 | 
						|
        return (int)r;
 | 
						|
      }
 | 
						|
      r = explode_lit8(__G__ tb, tl, td, bb, bl, bd);
 | 
						|
    }
 | 
						|
    else                                        /* else 4K */
 | 
						|
    {
 | 
						|
      if ((r = huft_build(__G__ l, 64, 0, cpdist4, extra, &td, &bd)) != 0)
 | 
						|
      {
 | 
						|
        if (r == 1)
 | 
						|
          huft_free(td);
 | 
						|
        huft_free(tl);
 | 
						|
        huft_free(tb);
 | 
						|
        return (int)r;
 | 
						|
      }
 | 
						|
      r = explode_lit4(__G__ tb, tl, td, bb, bl, bd);
 | 
						|
    }
 | 
						|
    huft_free(td);
 | 
						|
    huft_free(tl);
 | 
						|
    huft_free(tb);
 | 
						|
  }
 | 
						|
  else
 | 
						|
 | 
						|
 | 
						|
  /* No literal tree--minimum match length is 2 */
 | 
						|
  {
 | 
						|
    if ((r = get_tree(__G__ l, 64)) != 0)
 | 
						|
      return (int)r;
 | 
						|
    if ((r = huft_build(__G__ l, 64, 0, cplen2, extra, &tl, &bl)) != 0)
 | 
						|
    {
 | 
						|
      if (r == 1)
 | 
						|
        huft_free(tl);
 | 
						|
      return (int)r;
 | 
						|
    }
 | 
						|
    if ((r = get_tree(__G__ l, 64)) != 0)
 | 
						|
      return (int)r;
 | 
						|
    if (G.lrec.general_purpose_bit_flag & 2)      /* true if 8K */
 | 
						|
    {
 | 
						|
      if ((r = huft_build(__G__ l, 64, 0, cpdist8, extra, &td, &bd)) != 0)
 | 
						|
      {
 | 
						|
        if (r == 1)
 | 
						|
          huft_free(td);
 | 
						|
        huft_free(tl);
 | 
						|
        return (int)r;
 | 
						|
      }
 | 
						|
      r = explode_nolit8(__G__ tl, td, bl, bd);
 | 
						|
    }
 | 
						|
    else                                        /* else 4K */
 | 
						|
    {
 | 
						|
      if ((r = huft_build(__G__ l, 64, 0, cpdist4, extra, &td, &bd)) != 0)
 | 
						|
      {
 | 
						|
        if (r == 1)
 | 
						|
          huft_free(td);
 | 
						|
        huft_free(tl);
 | 
						|
        return (int)r;
 | 
						|
      }
 | 
						|
      r = explode_nolit4(__G__ tl, td, bl, bd);
 | 
						|
    }
 | 
						|
    huft_free(td);
 | 
						|
    huft_free(tl);
 | 
						|
  }
 | 
						|
  Trace((stderr, "<%u > ", G.hufts));
 | 
						|
  return (int)r;
 | 
						|
}
 | 
						|
 | 
						|
/* so explode.c and inflate.c can be compiled together into one object: */
 | 
						|
#undef NEXTBYTE
 | 
						|
#undef NEEDBITS
 | 
						|
#undef DUMPBITS
 |