]> git.saurik.com Git - wxWidgets.git/blob - utils/Install/packzip/explode.c
Bitmap updates
[wxWidgets.git] / utils / Install / packzip / explode.c
1 /* explode.c -- put in the public domain by Mark Adler
2 version c15, 6 July 1996 */
3
4
5 /* You can do whatever you like with this source file, though I would
6 prefer that if you modify it and redistribute it that you include
7 comments to that effect with your name and the date. Thank you.
8
9 History:
10 vers date who what
11 ---- --------- -------------- ------------------------------------
12 c1 30 Mar 92 M. Adler explode that uses huft_build from inflate
13 (this gives over a 70% speed improvement
14 over the original unimplode.c, which
15 decoded a bit at a time)
16 c2 4 Apr 92 M. Adler fixed bug for file sizes a multiple of 32k.
17 c3 10 Apr 92 M. Adler added a little memory tracking if DEBUG
18 c4 11 Apr 92 M. Adler added NOMEMCPY do kill use of memcpy()
19 c5 21 Apr 92 M. Adler added the WSIZE #define to allow reducing
20 the 32K window size for specialized
21 applications.
22 c6 31 May 92 M. Adler added typecasts to eliminate some warnings
23 c7 27 Jun 92 G. Roelofs added more typecasts.
24 c8 17 Oct 92 G. Roelofs changed ULONG/UWORD/byte to ulg/ush/uch.
25 c9 19 Jul 93 J. Bush added more typecasts (to return values);
26 made l[256] array static for Amiga.
27 c10 8 Oct 93 G. Roelofs added used_csize for diagnostics; added
28 buf and unshrink arguments to flush();
29 undef'd various macros at end for Turbo C;
30 removed NEXTBYTE macro (now in unzip.h)
31 and bytebuf variable (not used); changed
32 memset() to memzero().
33 c11 9 Jan 94 M. Adler fixed incorrect used_csize calculation.
34 c12 9 Apr 94 G. Roelofs fixed split comments on preprocessor lines
35 to avoid bug in Encore compiler.
36 c13 25 Aug 94 M. Adler fixed distance-length comment (orig c9 fix)
37 c14 22 Nov 95 S. Maxwell removed unnecessary "static" on auto array
38 c15 6 Jul 96 W. Haidinger added ulg typecasts to flush() calls.
39 c16 8 Feb 98 C. Spieler added ZCONST modifiers to const tables
40 and #ifdef DEBUG around debugging code.
41 c16b 25 Mar 98 C. Spieler modified DLL code for slide redirection.
42 */
43
44
45 /*
46 Explode imploded (PKZIP method 6 compressed) data. This compression
47 method searches for as much of the current string of bytes (up to a length
48 of ~320) in the previous 4K or 8K bytes. If it doesn't find any matches
49 (of at least length 2 or 3), it codes the next byte. Otherwise, it codes
50 the length of the matched string and its distance backwards from the
51 current position. Single bytes ("literals") are preceded by a one (a
52 single bit) and are either uncoded (the eight bits go directly into the
53 compressed stream for a total of nine bits) or Huffman coded with a
54 supplied literal code tree. If literals are coded, then the minimum match
55 length is three, otherwise it is two.
56
57 There are therefore four kinds of imploded streams: 8K search with coded
58 literals (min match = 3), 4K search with coded literals (min match = 3),
59 8K with uncoded literals (min match = 2), and 4K with uncoded literals
60 (min match = 2). The kind of stream is identified in two bits of a
61 general purpose bit flag that is outside of the compressed stream.
62
63 Distance-length pairs for matched strings are preceded by a zero bit (to
64 distinguish them from literals) and are always coded. The distance comes
65 first and is either the low six (4K) or low seven (8K) bits of the
66 distance (uncoded), followed by the high six bits of the distance coded.
67 Then the length is six bits coded (0..63 + min match length), and if the
68 maximum such length is coded, then it's followed by another eight bits
69 (uncoded) to be added to the coded length. This gives a match length
70 range of 2..320 or 3..321 bytes.
71
72 The literal, length, and distance codes are all represented in a slightly
73 compressed form themselves. What is sent are the lengths of the codes for
74 each value, which is sufficient to construct the codes. Each byte of the
75 code representation is the code length (the low four bits representing
76 1..16), and the number of values sequentially with that length (the high
77 four bits also representing 1..16). There are 256 literal code values (if
78 literals are coded), 64 length code values, and 64 distance code values,
79 in that order at the beginning of the compressed stream. Each set of code
80 values is preceded (redundantly) with a byte indicating how many bytes are
81 in the code description that follows, in the range 1..256.
82
83 The codes themselves are decoded using tables made by huft_build() from
84 the bit lengths. That routine and its comments are in the inflate.c
85 module.
86 */
87
88 #define UNZIP_INTERNAL
89 #include "unzip.h" /* must supply slide[] (uch) array and NEXTBYTE macro */
90
91 #ifndef WSIZE
92 # define WSIZE 0x8000 /* window size--must be a power of two, and */
93 #endif /* at least 8K for zip's implode method */
94
95 #if (defined(DLL) && !defined(NO_SLIDE_REDIR))
96 # define wsize G._wsize
97 #else
98 # define wsize WSIZE
99 #endif
100
101 /* routines here */
102 static int get_tree OF((__GPRO__ unsigned *l, unsigned n));
103 static int explode_lit8 OF((__GPRO__ struct huft *tb, struct huft *tl,
104 struct huft *td, int bb, int bl, int bd));
105 static int explode_lit4 OF((__GPRO__ struct huft *tb, struct huft *tl,
106 struct huft *td, int bb, int bl, int bd));
107 static int explode_nolit8 OF((__GPRO__ struct huft *tl, struct huft *td,
108 int bl, int bd));
109 static int explode_nolit4 OF((__GPRO__ struct huft *tl, struct huft *td,
110 int bl, int bd));
111 int explode OF((__GPRO));
112
113
114 /* The implode algorithm uses a sliding 4K or 8K byte window on the
115 uncompressed stream to find repeated byte strings. This is implemented
116 here as a circular buffer. The index is updated simply by incrementing
117 and then and'ing with 0x0fff (4K-1) or 0x1fff (8K-1). Here, the 32K
118 buffer of inflate is used, and it works just as well to always have
119 a 32K circular buffer, so the index is anded with 0x7fff. This is
120 done to allow the window to also be used as the output buffer. */
121 /* This must be supplied in an external module useable like "uch slide[8192];"
122 or "uch *slide;", where the latter would be malloc'ed. In unzip, slide[]
123 is actually a 32K area for use by inflate, which uses a 32K sliding window.
124 */
125
126
127 /* Tables for length and distance */
128 static ZCONST ush cplen2[] =
129 {2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17,
130 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34,
131 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51,
132 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65};
133 static ZCONST ush cplen3[] =
134 {3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18,
135 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35,
136 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52,
137 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66};
138 static ZCONST ush extra[] =
139 {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
140 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
141 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
142 8};
143 static ZCONST ush cpdist4[] =
144 {1, 65, 129, 193, 257, 321, 385, 449, 513, 577, 641, 705,
145 769, 833, 897, 961, 1025, 1089, 1153, 1217, 1281, 1345, 1409, 1473,
146 1537, 1601, 1665, 1729, 1793, 1857, 1921, 1985, 2049, 2113, 2177,
147 2241, 2305, 2369, 2433, 2497, 2561, 2625, 2689, 2753, 2817, 2881,
148 2945, 3009, 3073, 3137, 3201, 3265, 3329, 3393, 3457, 3521, 3585,
149 3649, 3713, 3777, 3841, 3905, 3969, 4033};
150 static ZCONST ush cpdist8[] =
151 {1, 129, 257, 385, 513, 641, 769, 897, 1025, 1153, 1281,
152 1409, 1537, 1665, 1793, 1921, 2049, 2177, 2305, 2433, 2561, 2689,
153 2817, 2945, 3073, 3201, 3329, 3457, 3585, 3713, 3841, 3969, 4097,
154 4225, 4353, 4481, 4609, 4737, 4865, 4993, 5121, 5249, 5377, 5505,
155 5633, 5761, 5889, 6017, 6145, 6273, 6401, 6529, 6657, 6785, 6913,
156 7041, 7169, 7297, 7425, 7553, 7681, 7809, 7937, 8065};
157
158
159 /* Macros for inflate() bit peeking and grabbing.
160 The usage is:
161
162 NEEDBITS(j)
163 x = b & mask_bits[j];
164 DUMPBITS(j)
165
166 where NEEDBITS makes sure that b has at least j bits in it, and
167 DUMPBITS removes the bits from b. The macros use the variable k
168 for the number of bits in b. Normally, b and k are register
169 variables for speed.
170 */
171
172 #define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE)<<k;k+=8;}}
173 #define DUMPBITS(n) {b>>=(n);k-=(n);}
174
175
176
177 static int get_tree(__G__ l, n)
178 __GDEF
179 unsigned *l; /* bit lengths */
180 unsigned n; /* number expected */
181 /* Get the bit lengths for a code representation from the compressed
182 stream. If get_tree() returns 4, then there is an error in the data.
183 Otherwise zero is returned. */
184 {
185 unsigned i; /* bytes remaining in list */
186 unsigned k; /* lengths entered */
187 unsigned j; /* number of codes */
188 unsigned b; /* bit length for those codes */
189
190
191 /* get bit lengths */
192 i = NEXTBYTE + 1; /* length/count pairs to read */
193 k = 0; /* next code */
194 do {
195 b = ((j = NEXTBYTE) & 0xf) + 1; /* bits in code (1..16) */
196 j = ((j & 0xf0) >> 4) + 1; /* codes with those bits (1..16) */
197 if (k + j > n)
198 return 4; /* don't overflow l[] */
199 do {
200 l[k++] = b;
201 } while (--j);
202 } while (--i);
203 return k != n ? 4 : 0; /* should have read n of them */
204 }
205
206
207
208 static int explode_lit8(__G__ tb, tl, td, bb, bl, bd)
209 __GDEF
210 struct huft *tb, *tl, *td; /* literal, length, and distance tables */
211 int bb, bl, bd; /* number of bits decoded by those */
212 /* Decompress the imploded data using coded literals and an 8K sliding
213 window. */
214 {
215 long s; /* bytes to decompress */
216 register unsigned e; /* table entry flag/number of extra bits */
217 unsigned n, d; /* length and index for copy */
218 unsigned w; /* current window position */
219 struct huft *t; /* pointer to table entry */
220 unsigned mb, ml, md; /* masks for bb, bl, and bd bits */
221 register ulg b; /* bit buffer */
222 register unsigned k; /* number of bits in bit buffer */
223 unsigned u; /* true if unflushed */
224
225
226 /* explode the coded data */
227 b = k = w = 0; /* initialize bit buffer, window */
228 u = 1; /* buffer unflushed */
229 mb = mask_bits[bb]; /* precompute masks for speed */
230 ml = mask_bits[bl];
231 md = mask_bits[bd];
232 s = G.ucsize;
233 while (s > 0) /* do until ucsize bytes uncompressed */
234 {
235 NEEDBITS(1)
236 if (b & 1) /* then literal--decode it */
237 {
238 DUMPBITS(1)
239 s--;
240 NEEDBITS((unsigned)bb) /* get coded literal */
241 if ((e = (t = tb + ((~(unsigned)b) & mb))->e) > 16)
242 do {
243 if (e == 99)
244 return 1;
245 DUMPBITS(t->b)
246 e -= 16;
247 NEEDBITS(e)
248 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
249 DUMPBITS(t->b)
250 redirSlide[w++] = (uch)t->v.n;
251 if (w == wsize)
252 {
253 flush(__G__ redirSlide, (ulg)w, 0);
254 w = u = 0;
255 }
256 }
257 else /* else distance/length */
258 {
259 DUMPBITS(1)
260 NEEDBITS(7) /* get distance low bits */
261 d = (unsigned)b & 0x7f;
262 DUMPBITS(7)
263 NEEDBITS((unsigned)bd) /* get coded distance high bits */
264 if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
265 do {
266 if (e == 99)
267 return 1;
268 DUMPBITS(t->b)
269 e -= 16;
270 NEEDBITS(e)
271 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
272 DUMPBITS(t->b)
273 d = w - d - t->v.n; /* construct offset */
274 NEEDBITS((unsigned)bl) /* get coded length */
275 if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
276 do {
277 if (e == 99)
278 return 1;
279 DUMPBITS(t->b)
280 e -= 16;
281 NEEDBITS(e)
282 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
283 DUMPBITS(t->b)
284 n = t->v.n;
285 if (e) /* get length extra bits */
286 {
287 NEEDBITS(8)
288 n += (unsigned)b & 0xff;
289 DUMPBITS(8)
290 }
291
292 /* do the copy */
293 s -= n;
294 do {
295 #if (defined(DLL) && !defined(NO_SLIDE_REDIR))
296 if (G.redirect_slide) {
297 /* &= w/ wsize not needed and wrong if redirect */
298 if (d >= wsize)
299 return 1;
300 n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
301 } else
302 #endif
303 n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
304 if (u && w <= d)
305 {
306 memzero(redirSlide + w, e);
307 w += e;
308 d += e;
309 }
310 else
311 #ifndef NOMEMCPY
312 if (w - d >= e) /* (this test assumes unsigned comparison) */
313 {
314 memcpy(redirSlide + w, redirSlide + d, e);
315 w += e;
316 d += e;
317 }
318 else /* do it slow to avoid memcpy() overlap */
319 #endif /* !NOMEMCPY */
320 do {
321 redirSlide[w++] = redirSlide[d++];
322 } while (--e);
323 if (w == wsize)
324 {
325 flush(__G__ redirSlide, (ulg)w, 0);
326 w = u = 0;
327 }
328 } while (n);
329 }
330 }
331
332 /* flush out redirSlide */
333 flush(__G__ redirSlide, (ulg)w, 0);
334 if (G.csize + G.incnt + (k >> 3)) /* should have read csize bytes, but */
335 { /* sometimes read one too many: k>>3 compensates */
336 G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
337 return 5;
338 }
339 return 0;
340 }
341
342
343
344 static int explode_lit4(__G__ tb, tl, td, bb, bl, bd)
345 __GDEF
346 struct huft *tb, *tl, *td; /* literal, length, and distance tables */
347 int bb, bl, bd; /* number of bits decoded by those */
348 /* Decompress the imploded data using coded literals and a 4K sliding
349 window. */
350 {
351 long s; /* bytes to decompress */
352 register unsigned e; /* table entry flag/number of extra bits */
353 unsigned n, d; /* length and index for copy */
354 unsigned w; /* current window position */
355 struct huft *t; /* pointer to table entry */
356 unsigned mb, ml, md; /* masks for bb, bl, and bd bits */
357 register ulg b; /* bit buffer */
358 register unsigned k; /* number of bits in bit buffer */
359 unsigned u; /* true if unflushed */
360
361
362 /* explode the coded data */
363 b = k = w = 0; /* initialize bit buffer, window */
364 u = 1; /* buffer unflushed */
365 mb = mask_bits[bb]; /* precompute masks for speed */
366 ml = mask_bits[bl];
367 md = mask_bits[bd];
368 s = G.ucsize;
369 while (s > 0) /* do until ucsize bytes uncompressed */
370 {
371 NEEDBITS(1)
372 if (b & 1) /* then literal--decode it */
373 {
374 DUMPBITS(1)
375 s--;
376 NEEDBITS((unsigned)bb) /* get coded literal */
377 if ((e = (t = tb + ((~(unsigned)b) & mb))->e) > 16)
378 do {
379 if (e == 99)
380 return 1;
381 DUMPBITS(t->b)
382 e -= 16;
383 NEEDBITS(e)
384 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
385 DUMPBITS(t->b)
386 redirSlide[w++] = (uch)t->v.n;
387 if (w == wsize)
388 {
389 flush(__G__ redirSlide, (ulg)w, 0);
390 w = u = 0;
391 }
392 }
393 else /* else distance/length */
394 {
395 DUMPBITS(1)
396 NEEDBITS(6) /* get distance low bits */
397 d = (unsigned)b & 0x3f;
398 DUMPBITS(6)
399 NEEDBITS((unsigned)bd) /* get coded distance high bits */
400 if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
401 do {
402 if (e == 99)
403 return 1;
404 DUMPBITS(t->b)
405 e -= 16;
406 NEEDBITS(e)
407 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
408 DUMPBITS(t->b)
409 d = w - d - t->v.n; /* construct offset */
410 NEEDBITS((unsigned)bl) /* get coded length */
411 if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
412 do {
413 if (e == 99)
414 return 1;
415 DUMPBITS(t->b)
416 e -= 16;
417 NEEDBITS(e)
418 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
419 DUMPBITS(t->b)
420 n = t->v.n;
421 if (e) /* get length extra bits */
422 {
423 NEEDBITS(8)
424 n += (unsigned)b & 0xff;
425 DUMPBITS(8)
426 }
427
428 /* do the copy */
429 s -= n;
430 do {
431 #if (defined(DLL) && !defined(NO_SLIDE_REDIR))
432 if (G.redirect_slide) {
433 /* &= w/ wsize not needed and wrong if redirect */
434 if (d >= wsize)
435 return 1;
436 n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
437 } else
438 #endif
439 n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
440 if (u && w <= d)
441 {
442 memzero(redirSlide + w, e);
443 w += e;
444 d += e;
445 }
446 else
447 #ifndef NOMEMCPY
448 if (w - d >= e) /* (this test assumes unsigned comparison) */
449 {
450 memcpy(redirSlide + w, redirSlide + d, e);
451 w += e;
452 d += e;
453 }
454 else /* do it slow to avoid memcpy() overlap */
455 #endif /* !NOMEMCPY */
456 do {
457 redirSlide[w++] = redirSlide[d++];
458 } while (--e);
459 if (w == wsize)
460 {
461 flush(__G__ redirSlide, (ulg)w, 0);
462 w = u = 0;
463 }
464 } while (n);
465 }
466 }
467
468 /* flush out redirSlide */
469 flush(__G__ redirSlide, (ulg)w, 0);
470 if (G.csize + G.incnt + (k >> 3)) /* should have read csize bytes, but */
471 { /* sometimes read one too many: k>>3 compensates */
472 G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
473 return 5;
474 }
475 return 0;
476 }
477
478
479
480 static int explode_nolit8(__G__ tl, td, bl, bd)
481 __GDEF
482 struct huft *tl, *td; /* length and distance decoder tables */
483 int bl, bd; /* number of bits decoded by tl[] and td[] */
484 /* Decompress the imploded data using uncoded literals and an 8K sliding
485 window. */
486 {
487 long s; /* bytes to decompress */
488 register unsigned e; /* table entry flag/number of extra bits */
489 unsigned n, d; /* length and index for copy */
490 unsigned w; /* current window position */
491 struct huft *t; /* pointer to table entry */
492 unsigned ml, md; /* masks for bl and bd bits */
493 register ulg b; /* bit buffer */
494 register unsigned k; /* number of bits in bit buffer */
495 unsigned u; /* true if unflushed */
496
497
498 /* explode the coded data */
499 b = k = w = 0; /* initialize bit buffer, window */
500 u = 1; /* buffer unflushed */
501 ml = mask_bits[bl]; /* precompute masks for speed */
502 md = mask_bits[bd];
503 s = G.ucsize;
504 while (s > 0) /* do until ucsize bytes uncompressed */
505 {
506 NEEDBITS(1)
507 if (b & 1) /* then literal--get eight bits */
508 {
509 DUMPBITS(1)
510 s--;
511 NEEDBITS(8)
512 redirSlide[w++] = (uch)b;
513 if (w == wsize)
514 {
515 flush(__G__ redirSlide, (ulg)w, 0);
516 w = u = 0;
517 }
518 DUMPBITS(8)
519 }
520 else /* else distance/length */
521 {
522 DUMPBITS(1)
523 NEEDBITS(7) /* get distance low bits */
524 d = (unsigned)b & 0x7f;
525 DUMPBITS(7)
526 NEEDBITS((unsigned)bd) /* get coded distance high bits */
527 if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
528 do {
529 if (e == 99)
530 return 1;
531 DUMPBITS(t->b)
532 e -= 16;
533 NEEDBITS(e)
534 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
535 DUMPBITS(t->b)
536 d = w - d - t->v.n; /* construct offset */
537 NEEDBITS((unsigned)bl) /* get coded length */
538 if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
539 do {
540 if (e == 99)
541 return 1;
542 DUMPBITS(t->b)
543 e -= 16;
544 NEEDBITS(e)
545 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
546 DUMPBITS(t->b)
547 n = t->v.n;
548 if (e) /* get length extra bits */
549 {
550 NEEDBITS(8)
551 n += (unsigned)b & 0xff;
552 DUMPBITS(8)
553 }
554
555 /* do the copy */
556 s -= n;
557 do {
558 #if (defined(DLL) && !defined(NO_SLIDE_REDIR))
559 if (G.redirect_slide) {
560 /* &= w/ wsize not needed and wrong if redirect */
561 if (d >= wsize)
562 return 1;
563 n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
564 } else
565 #endif
566 n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
567 if (u && w <= d)
568 {
569 memzero(redirSlide + w, e);
570 w += e;
571 d += e;
572 }
573 else
574 #ifndef NOMEMCPY
575 if (w - d >= e) /* (this test assumes unsigned comparison) */
576 {
577 memcpy(redirSlide + w, redirSlide + d, e);
578 w += e;
579 d += e;
580 }
581 else /* do it slow to avoid memcpy() overlap */
582 #endif /* !NOMEMCPY */
583 do {
584 redirSlide[w++] = redirSlide[d++];
585 } while (--e);
586 if (w == wsize)
587 {
588 flush(__G__ redirSlide, (ulg)w, 0);
589 w = u = 0;
590 }
591 } while (n);
592 }
593 }
594
595 /* flush out redirSlide */
596 flush(__G__ redirSlide, (ulg)w, 0);
597 if (G.csize + G.incnt + (k >> 3)) /* should have read csize bytes, but */
598 { /* sometimes read one too many: k>>3 compensates */
599 G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
600 return 5;
601 }
602 return 0;
603 }
604
605
606
607 static int explode_nolit4(__G__ tl, td, bl, bd)
608 __GDEF
609 struct huft *tl, *td; /* length and distance decoder tables */
610 int bl, bd; /* number of bits decoded by tl[] and td[] */
611 /* Decompress the imploded data using uncoded literals and a 4K sliding
612 window. */
613 {
614 long s; /* bytes to decompress */
615 register unsigned e; /* table entry flag/number of extra bits */
616 unsigned n, d; /* length and index for copy */
617 unsigned w; /* current window position */
618 struct huft *t; /* pointer to table entry */
619 unsigned ml, md; /* masks for bl and bd bits */
620 register ulg b; /* bit buffer */
621 register unsigned k; /* number of bits in bit buffer */
622 unsigned u; /* true if unflushed */
623
624
625 /* explode the coded data */
626 b = k = w = 0; /* initialize bit buffer, window */
627 u = 1; /* buffer unflushed */
628 ml = mask_bits[bl]; /* precompute masks for speed */
629 md = mask_bits[bd];
630 s = G.ucsize;
631 while (s > 0) /* do until ucsize bytes uncompressed */
632 {
633 NEEDBITS(1)
634 if (b & 1) /* then literal--get eight bits */
635 {
636 DUMPBITS(1)
637 s--;
638 NEEDBITS(8)
639 redirSlide[w++] = (uch)b;
640 if (w == wsize)
641 {
642 flush(__G__ redirSlide, (ulg)w, 0);
643 w = u = 0;
644 }
645 DUMPBITS(8)
646 }
647 else /* else distance/length */
648 {
649 DUMPBITS(1)
650 NEEDBITS(6) /* get distance low bits */
651 d = (unsigned)b & 0x3f;
652 DUMPBITS(6)
653 NEEDBITS((unsigned)bd) /* get coded distance high bits */
654 if ((e = (t = td + ((~(unsigned)b) & md))->e) > 16)
655 do {
656 if (e == 99)
657 return 1;
658 DUMPBITS(t->b)
659 e -= 16;
660 NEEDBITS(e)
661 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
662 DUMPBITS(t->b)
663 d = w - d - t->v.n; /* construct offset */
664 NEEDBITS((unsigned)bl) /* get coded length */
665 if ((e = (t = tl + ((~(unsigned)b) & ml))->e) > 16)
666 do {
667 if (e == 99)
668 return 1;
669 DUMPBITS(t->b)
670 e -= 16;
671 NEEDBITS(e)
672 } while ((e = (t = t->v.t + ((~(unsigned)b) & mask_bits[e]))->e) > 16);
673 DUMPBITS(t->b)
674 n = t->v.n;
675 if (e) /* get length extra bits */
676 {
677 NEEDBITS(8)
678 n += (unsigned)b & 0xff;
679 DUMPBITS(8)
680 }
681
682 /* do the copy */
683 s -= n;
684 do {
685 #if (defined(DLL) && !defined(NO_SLIDE_REDIR))
686 if (G.redirect_slide) {
687 /* &= w/ wsize not needed and wrong if redirect */
688 if (d >= wsize)
689 return 1;
690 n -= (e = (e = wsize - (d > w ? d : w)) > n ? n : e);
691 } else
692 #endif
693 n -= (e = (e = wsize - ((d &= wsize-1) > w ? d : w)) > n ? n : e);
694 if (u && w <= d)
695 {
696 memzero(redirSlide + w, e);
697 w += e;
698 d += e;
699 }
700 else
701 #ifndef NOMEMCPY
702 if (w - d >= e) /* (this test assumes unsigned comparison) */
703 {
704 memcpy(redirSlide + w, redirSlide + d, e);
705 w += e;
706 d += e;
707 }
708 else /* do it slow to avoid memcpy() overlap */
709 #endif /* !NOMEMCPY */
710 do {
711 redirSlide[w++] = redirSlide[d++];
712 } while (--e);
713 if (w == wsize)
714 {
715 flush(__G__ redirSlide, (ulg)w, 0);
716 w = u = 0;
717 }
718 } while (n);
719 }
720 }
721
722 /* flush out redirSlide */
723 flush(__G__ redirSlide, (ulg)w, 0);
724 if (G.csize + G.incnt + (k >> 3)) /* should have read csize bytes, but */
725 { /* sometimes read one too many: k>>3 compensates */
726 G.used_csize = G.lrec.csize - G.csize - G.incnt - (k >> 3);
727 return 5;
728 }
729 return 0;
730 }
731
732
733
734 int explode(__G)
735 __GDEF
736 /* Explode an imploded compressed stream. Based on the general purpose
737 bit flag, decide on coded or uncoded literals, and an 8K or 4K sliding
738 window. Construct the literal (if any), length, and distance codes and
739 the tables needed to decode them (using huft_build() from inflate.c),
740 and call the appropriate routine for the type of data in the remainder
741 of the stream. The four routines are nearly identical, differing only
742 in whether the literal is decoded or simply read in, and in how many
743 bits are read in, uncoded, for the low distance bits. */
744 {
745 unsigned r; /* return codes */
746 struct huft *tb; /* literal code table */
747 struct huft *tl; /* length code table */
748 struct huft *td; /* distance code table */
749 int bb; /* bits for tb */
750 int bl; /* bits for tl */
751 int bd; /* bits for td */
752 unsigned l[256]; /* bit lengths for codes */
753
754 #if (defined(DLL) && !defined(NO_SLIDE_REDIR))
755 if (G.redirect_slide)
756 wsize = G.redirect_size, redirSlide = G.redirect_buffer;
757 else
758 wsize = WSIZE, redirSlide = slide;
759 #endif
760
761 /* Tune base table sizes. Note: I thought that to truly optimize speed,
762 I would have to select different bl, bd, and bb values for different
763 compressed file sizes. I was surprised to find out that the values of
764 7, 7, and 9 worked best over a very wide range of sizes, except that
765 bd = 8 worked marginally better for large compressed sizes. */
766 bl = 7;
767 bd = (G.csize + G.incnt) > 200000L ? 8 : 7;
768
769
770 /* With literal tree--minimum match length is 3 */
771 #ifdef DEBUG
772 G.hufts = 0; /* initialize huft's malloc'ed */
773 #endif
774 if (G.lrec.general_purpose_bit_flag & 4)
775 {
776 bb = 9; /* base table size for literals */
777 if ((r = get_tree(__G__ l, 256)) != 0)
778 return (int)r;
779 if ((r = huft_build(__G__ l, 256, 256, NULL, NULL, &tb, &bb)) != 0)
780 {
781 if (r == 1)
782 huft_free(tb);
783 return (int)r;
784 }
785 if ((r = get_tree(__G__ l, 64)) != 0)
786 return (int)r;
787 if ((r = huft_build(__G__ l, 64, 0, cplen3, extra, &tl, &bl)) != 0)
788 {
789 if (r == 1)
790 huft_free(tl);
791 huft_free(tb);
792 return (int)r;
793 }
794 if ((r = get_tree(__G__ l, 64)) != 0)
795 return (int)r;
796 if (G.lrec.general_purpose_bit_flag & 2) /* true if 8K */
797 {
798 if ((r = huft_build(__G__ l, 64, 0, cpdist8, extra, &td, &bd)) != 0)
799 {
800 if (r == 1)
801 huft_free(td);
802 huft_free(tl);
803 huft_free(tb);
804 return (int)r;
805 }
806 r = explode_lit8(__G__ tb, tl, td, bb, bl, bd);
807 }
808 else /* else 4K */
809 {
810 if ((r = huft_build(__G__ l, 64, 0, cpdist4, extra, &td, &bd)) != 0)
811 {
812 if (r == 1)
813 huft_free(td);
814 huft_free(tl);
815 huft_free(tb);
816 return (int)r;
817 }
818 r = explode_lit4(__G__ tb, tl, td, bb, bl, bd);
819 }
820 huft_free(td);
821 huft_free(tl);
822 huft_free(tb);
823 }
824 else
825
826
827 /* No literal tree--minimum match length is 2 */
828 {
829 if ((r = get_tree(__G__ l, 64)) != 0)
830 return (int)r;
831 if ((r = huft_build(__G__ l, 64, 0, cplen2, extra, &tl, &bl)) != 0)
832 {
833 if (r == 1)
834 huft_free(tl);
835 return (int)r;
836 }
837 if ((r = get_tree(__G__ l, 64)) != 0)
838 return (int)r;
839 if (G.lrec.general_purpose_bit_flag & 2) /* true if 8K */
840 {
841 if ((r = huft_build(__G__ l, 64, 0, cpdist8, extra, &td, &bd)) != 0)
842 {
843 if (r == 1)
844 huft_free(td);
845 huft_free(tl);
846 return (int)r;
847 }
848 r = explode_nolit8(__G__ tl, td, bl, bd);
849 }
850 else /* else 4K */
851 {
852 if ((r = huft_build(__G__ l, 64, 0, cpdist4, extra, &td, &bd)) != 0)
853 {
854 if (r == 1)
855 huft_free(td);
856 huft_free(tl);
857 return (int)r;
858 }
859 r = explode_nolit4(__G__ tl, td, bl, bd);
860 }
861 huft_free(td);
862 huft_free(tl);
863 }
864 Trace((stderr, "<%u > ", G.hufts));
865 return (int)r;
866 }
867
868 /* so explode.c and inflate.c can be compiled together into one object: */
869 #undef NEXTBYTE
870 #undef NEEDBITS
871 #undef DUMPBITS