All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.commons.compress.archivers.zip.ZipArchiveInputStream Maven / Gradle / Ivy

Go to download

Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4, Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.

There is a newer version: 1.26.1
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */
package org.apache.commons.compress.archivers.zip;

import static org.apache.commons.compress.archivers.zip.ZipConstants.DWORD;
import static org.apache.commons.compress.archivers.zip.ZipConstants.SHORT;
import static org.apache.commons.compress.archivers.zip.ZipConstants.WORD;
import static org.apache.commons.compress.archivers.zip.ZipConstants.ZIP64_MAGIC;

import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.EOFException;
import java.io.IOException;
import java.io.InputStream;
import java.io.PushbackInputStream;
import java.math.BigInteger;
import java.nio.ByteBuffer;
import java.util.Arrays;
import java.util.Objects;
import java.util.zip.CRC32;
import java.util.zip.DataFormatException;
import java.util.zip.Inflater;
import java.util.zip.ZipEntry;
import java.util.zip.ZipException;

import org.apache.commons.compress.archivers.ArchiveEntry;
import org.apache.commons.compress.archivers.ArchiveInputStream;
import org.apache.commons.compress.compressors.bzip2.BZip2CompressorInputStream;
import org.apache.commons.compress.compressors.deflate64.Deflate64CompressorInputStream;
import org.apache.commons.compress.utils.ArchiveUtils;
import org.apache.commons.compress.utils.IOUtils;
import org.apache.commons.compress.utils.InputStreamStatistics;

/**
 * Implements an input stream that can read Zip archives.
 *
 * 

As of Apache Commons Compress it transparently supports Zip64 * extensions and thus individual entries and archives larger than 4 * GB or with more than 65536 entries.

* *

The {@link ZipFile} class is preferred when reading from files * as {@link ZipArchiveInputStream} is limited by not being able to * read the central directory header before returning entries. In * particular {@link ZipArchiveInputStream}

* *
    * *
  • may return entries that are not part of the central directory * at all and shouldn't be considered part of the archive.
  • * *
  • may return several entries with the same name.
  • * *
  • will not return internal or external attributes.
  • * *
  • may return incomplete extra field data.
  • * *
  • may return unknown sizes and CRC values for entries until the * next entry has been reached if the archive uses the data * descriptor feature.
  • * *
* * @see ZipFile * @NotThreadSafe */ public class ZipArchiveInputStream extends ArchiveInputStream implements InputStreamStatistics { /** * Bounded input stream adapted from commons-io */ private class BoundedInputStream extends InputStream { /** the wrapped input stream */ private final InputStream in; /** the max length to provide */ private final long max; /** the number of bytes already returned */ private long pos; /** * Creates a new {@code BoundedInputStream} that wraps the given input * stream and limits it to a certain size. * * @param in The wrapped input stream * @param size The maximum number of bytes to return */ public BoundedInputStream(final InputStream in, final long size) { this.max = size; this.in = in; } @Override public int available() throws IOException { if (max >= 0 && pos >= max) { return 0; } return in.available(); } @Override public int read() throws IOException { if (max >= 0 && pos >= max) { return -1; } final int result = in.read(); pos++; count(1); current.bytesReadFromStream++; return result; } @Override public int read(final byte[] b) throws IOException { return this.read(b, 0, b.length); } @Override public int read(final byte[] b, final int off, final int len) throws IOException { if (len == 0) { return 0; } if (max >= 0 && pos >= max) { return -1; } final long maxRead = max >= 0 ? Math.min(len, max - pos) : len; final int bytesRead = in.read(b, off, (int) maxRead); if (bytesRead == -1) { return -1; } pos += bytesRead; count(bytesRead); current.bytesReadFromStream += bytesRead; return bytesRead; } @Override public long skip(final long n) throws IOException { final long toSkip = max >= 0 ? Math.min(n, max - pos) : n; final long skippedBytes = IOUtils.skip(in, toSkip); pos += skippedBytes; return skippedBytes; } } /** * Structure collecting information for the entry that is * currently being read. */ private static final class CurrentEntry { /** * Current ZIP entry. */ private final ZipArchiveEntry entry = new ZipArchiveEntry(); /** * Does the entry use a data descriptor? */ private boolean hasDataDescriptor; /** * Does the entry have a ZIP64 extended information extra field. */ private boolean usesZip64; /** * Number of bytes of entry content read by the client if the * entry is STORED. */ private long bytesRead; /** * Number of bytes of entry content read from the stream. * *

This may be more than the actual entry's length as some * stuff gets buffered up and needs to be pushed back when the * end of the entry has been reached.

*/ private long bytesReadFromStream; /** * The checksum calculated as the current entry is read. */ private final CRC32 crc = new CRC32(); /** * The input stream decompressing the data for shrunk and imploded entries. */ private InputStream inputStream; @SuppressWarnings("unchecked") // Caller beware private T checkInputStream() { return (T) Objects.requireNonNull(inputStream, "inputStream"); } } private static final int LFH_LEN = 30; /* local file header signature WORD version needed to extract SHORT general purpose bit flag SHORT compression method SHORT last mod file time SHORT last mod file date SHORT crc-32 WORD compressed size WORD uncompressed size WORD file name length SHORT extra field length SHORT */ private static final int CFH_LEN = 46; /* central file header signature WORD version made by SHORT version needed to extract SHORT general purpose bit flag SHORT compression method SHORT last mod file time SHORT last mod file date SHORT crc-32 WORD compressed size WORD uncompressed size WORD file name length SHORT extra field length SHORT file comment length SHORT disk number start SHORT internal file attributes SHORT external file attributes WORD relative offset of local header WORD */ private static final long TWO_EXP_32 = ZIP64_MAGIC + 1; private static final String USE_ZIPFILE_INSTEAD_OF_STREAM_DISCLAIMER = " while reading a stored entry using data descriptor. Either the archive is broken" + " or it can not be read using ZipArchiveInputStream and you must use ZipFile." + " A common cause for this is a ZIP archive containing a ZIP archive." + " See http://commons.apache.org/proper/commons-compress/zip.html#ZipArchiveInputStream_vs_ZipFile"; private static final byte[] LFH = ZipLong.LFH_SIG.getBytes(); private static final byte[] CFH = ZipLong.CFH_SIG.getBytes(); private static final byte[] DD = ZipLong.DD_SIG.getBytes(); private static final byte[] APK_SIGNING_BLOCK_MAGIC = { 'A', 'P', 'K', ' ', 'S', 'i', 'g', ' ', 'B', 'l', 'o', 'c', 'k', ' ', '4', '2', }; private static final BigInteger LONG_MAX = BigInteger.valueOf(Long.MAX_VALUE); private static boolean checksig(final byte[] signature, final byte[] expected) { for (int i = 0; i < expected.length; i++) { if (signature[i] != expected[i]) { return false; } } return true; } /** * Checks if the signature matches what is expected for a ZIP file. * Does not currently handle self-extracting ZIPs which may have arbitrary * leading content. * * @param signature the bytes to check * @param length the number of bytes to check * @return true, if this stream is a ZIP archive stream, false otherwise */ public static boolean matches(final byte[] signature, final int length) { if (length < ZipArchiveOutputStream.LFH_SIG.length) { return false; } return checksig(signature, ZipArchiveOutputStream.LFH_SIG) // normal file || checksig(signature, ZipArchiveOutputStream.EOCD_SIG) // empty zip || checksig(signature, ZipArchiveOutputStream.DD_SIG) // split zip || checksig(signature, ZipLong.SINGLE_SEGMENT_SPLIT_MARKER.getBytes()); } /** The ZIP encoding to use for file names and the file comment. */ private final ZipEncoding zipEncoding; // the provided encoding (for unit tests) final String encoding; /** Whether to look for and use Unicode extra fields. */ private final boolean useUnicodeExtraFields; /** Wrapped stream, will always be a PushbackInputStream. */ private final InputStream inputStream; /** Inflater used for all deflated entries. */ private final Inflater inf = new Inflater(true); /** Buffer used to read from the wrapped stream. */ private final ByteBuffer buf = ByteBuffer.allocate(ZipArchiveOutputStream.BUFFER_SIZE); /** The entry that is currently being read. */ private CurrentEntry current; /** Whether the stream has been closed. */ private boolean closed; /** Whether the stream has reached the central directory - and thus found all entries. */ private boolean hitCentralDirectory; /** * When reading a stored entry that uses the data descriptor this * stream has to read the full entry and caches it. This is the * cache. */ private ByteArrayInputStream lastStoredEntry; /** * Whether the stream will try to read STORED entries that use a data descriptor. * Setting it to true means we will not stop reading a entry with the compressed * size, instead we will stoping reading a entry when a data descriptor is met(by * finding the Data Descriptor Signature). This will completely break down in some * cases - like JARs in WARs. *

* See also : * https://issues.apache.org/jira/projects/COMPRESS/issues/COMPRESS-555 * https://github.com/apache/commons-compress/pull/137#issuecomment-690835644 */ private final boolean allowStoredEntriesWithDataDescriptor; /** Count decompressed bytes for current entry */ private long uncompressedCount; /** Whether the stream will try to skip the ZIP split signature(08074B50) at the beginning **/ private final boolean skipSplitSig; // cached buffers - must only be used locally in the class (COMPRESS-172 - reduce garbage collection) private final byte[] lfhBuf = new byte[LFH_LEN]; private final byte[] skipBuf = new byte[1024]; private final byte[] shortBuf = new byte[SHORT]; private final byte[] wordBuf = new byte[WORD]; private final byte[] twoDwordBuf = new byte[2 * DWORD]; private int entriesRead; /** * Create an instance using UTF-8 encoding * @param inputStream the stream to wrap */ public ZipArchiveInputStream(final InputStream inputStream) { this(inputStream, ZipEncodingHelper.UTF8); } /** * Create an instance using the specified encoding * @param inputStream the stream to wrap * @param encoding the encoding to use for file names, use null * for the platform's default encoding * @since 1.5 */ public ZipArchiveInputStream(final InputStream inputStream, final String encoding) { this(inputStream, encoding, true); } /** * Create an instance using the specified encoding * @param inputStream the stream to wrap * @param encoding the encoding to use for file names, use null * for the platform's default encoding * @param useUnicodeExtraFields whether to use InfoZIP Unicode * Extra Fields (if present) to set the file names. */ public ZipArchiveInputStream(final InputStream inputStream, final String encoding, final boolean useUnicodeExtraFields) { this(inputStream, encoding, useUnicodeExtraFields, false); } /** * Create an instance using the specified encoding * @param inputStream the stream to wrap * @param encoding the encoding to use for file names, use null * for the platform's default encoding * @param useUnicodeExtraFields whether to use InfoZIP Unicode * Extra Fields (if present) to set the file names. * @param allowStoredEntriesWithDataDescriptor whether the stream * will try to read STORED entries that use a data descriptor * @since 1.1 */ public ZipArchiveInputStream(final InputStream inputStream, final String encoding, final boolean useUnicodeExtraFields, final boolean allowStoredEntriesWithDataDescriptor) { this(inputStream, encoding, useUnicodeExtraFields, allowStoredEntriesWithDataDescriptor, false); } /** * Create an instance using the specified encoding * @param inputStream the stream to wrap * @param encoding the encoding to use for file names, use null * for the platform's default encoding * @param useUnicodeExtraFields whether to use InfoZIP Unicode * Extra Fields (if present) to set the file names. * @param allowStoredEntriesWithDataDescriptor whether the stream * will try to read STORED entries that use a data descriptor * @param skipSplitSig Whether the stream will try to skip the zip * split signature(08074B50) at the beginning. You will need to * set this to true if you want to read a split archive. * @since 1.20 */ public ZipArchiveInputStream(final InputStream inputStream, final String encoding, final boolean useUnicodeExtraFields, final boolean allowStoredEntriesWithDataDescriptor, final boolean skipSplitSig) { this.encoding = encoding; zipEncoding = ZipEncodingHelper.getZipEncoding(encoding); this.useUnicodeExtraFields = useUnicodeExtraFields; this.inputStream = new PushbackInputStream(inputStream, buf.capacity()); this.allowStoredEntriesWithDataDescriptor = allowStoredEntriesWithDataDescriptor; this.skipSplitSig = skipSplitSig; // haven't read anything so far buf.limit(0); } /** * Checks whether the current buffer contains the signature of a * "data descriptor", "local file header" or * "central directory entry". * *

If it contains such a signature, reads the data descriptor * and positions the stream right after the data descriptor.

*/ private boolean bufferContainsSignature(final ByteArrayOutputStream bos, final int offset, final int lastRead, final int expectedDDLen) throws IOException { boolean done = false; for (int i = 0; !done && i < offset + lastRead - 4; i++) { if (buf.array()[i] == LFH[0] && buf.array()[i + 1] == LFH[1]) { int expectDDPos = i; if (i >= expectedDDLen && (buf.array()[i + 2] == LFH[2] && buf.array()[i + 3] == LFH[3]) || (buf.array()[i + 2] == CFH[2] && buf.array()[i + 3] == CFH[3])) { // found a LFH or CFH: expectDDPos = i - expectedDDLen; done = true; } else if (buf.array()[i + 2] == DD[2] && buf.array()[i + 3] == DD[3]) { // found DD: done = true; } if (done) { // * push back bytes read in excess as well as the data // descriptor // * copy the remaining bytes to cache // * read data descriptor pushback(buf.array(), expectDDPos, offset + lastRead - expectDDPos); bos.write(buf.array(), 0, expectDDPos); readDataDescriptor(); } } } return done; } /** * If the last read bytes could hold a data descriptor and an * incomplete signature then save the last bytes to the front of * the buffer and cache everything in front of the potential data * descriptor into the given ByteArrayOutputStream. * *

Data descriptor plus incomplete signature (3 bytes in the * worst case) can be 20 bytes max.

*/ private int cacheBytesRead(final ByteArrayOutputStream bos, int offset, final int lastRead, final int expecteDDLen) { final int cacheable = offset + lastRead - expecteDDLen - 3; if (cacheable > 0) { bos.write(buf.array(), 0, cacheable); System.arraycopy(buf.array(), cacheable, buf.array(), 0, expecteDDLen + 3); offset = expecteDDLen + 3; } else { offset += lastRead; } return offset; } /** * Whether this class is able to read the given entry. * *

May return false if it is set up to use encryption or a * compression method that hasn't been implemented yet.

* @since 1.1 */ @Override public boolean canReadEntryData(final ArchiveEntry ae) { if (ae instanceof ZipArchiveEntry) { final ZipArchiveEntry ze = (ZipArchiveEntry) ae; return ZipUtil.canHandleEntryData(ze) && supportsDataDescriptorFor(ze) && supportsCompressedSizeFor(ze); } return false; } @Override public void close() throws IOException { if (!closed) { closed = true; try { inputStream.close(); } finally { inf.end(); } } } /** * Closes the current ZIP archive entry and positions the underlying * stream to the beginning of the next entry. All per-entry variables * and data structures are cleared. *

* If the compressed size of this entry is included in the entry header, * then any outstanding bytes are simply skipped from the underlying * stream without uncompressing them. This allows an entry to be safely * closed even if the compression method is unsupported. *

* In case we don't know the compressed size of this entry or have * already buffered too much data from the underlying stream to support * uncompression, then the uncompression process is completed and the * end position of the stream is adjusted based on the result of that * process. * * @throws IOException if an error occurs */ private void closeEntry() throws IOException { if (closed) { throw new IOException("The stream is closed"); } if (current == null) { return; } // Ensure all entry bytes are read if (currentEntryHasOutstandingBytes()) { drainCurrentEntryData(); } else { // this is guaranteed to exhaust the stream skip(Long.MAX_VALUE); //NOSONAR final long inB = current.entry.getMethod() == ZipArchiveOutputStream.DEFLATED ? getBytesInflated() : current.bytesRead; // this is at most a single read() operation and can't // exceed the range of int final int diff = (int) (current.bytesReadFromStream - inB); // Pushback any required bytes if (diff > 0) { pushback(buf.array(), buf.limit() - diff, diff); current.bytesReadFromStream -= diff; } // Drain remainder of entry if not all data bytes were required if (currentEntryHasOutstandingBytes()) { drainCurrentEntryData(); } } if (lastStoredEntry == null && current.hasDataDescriptor) { readDataDescriptor(); } inf.reset(); buf.clear().flip(); current = null; lastStoredEntry = null; } /** * If the compressed size of the current entry is included in the entry header * and there are any outstanding bytes in the underlying stream, then * this returns true. * * @return true, if current entry is determined to have outstanding bytes, false otherwise */ private boolean currentEntryHasOutstandingBytes() { return current.bytesReadFromStream <= current.entry.getCompressedSize() && !current.hasDataDescriptor; } /** * Read all data of the current entry from the underlying stream * that hasn't been read, yet. */ private void drainCurrentEntryData() throws IOException { long remaining = current.entry.getCompressedSize() - current.bytesReadFromStream; while (remaining > 0) { final long n = inputStream.read(buf.array(), 0, (int) Math.min(buf.capacity(), remaining)); if (n < 0) { throw new EOFException("Truncated ZIP entry: " + ArchiveUtils.sanitize(current.entry.getName())); } count(n); remaining -= n; } } private int fill() throws IOException { if (closed) { throw new IOException("The stream is closed"); } final int length = inputStream.read(buf.array()); if (length > 0) { buf.limit(length); count(buf.limit()); inf.setInput(buf.array(), 0, buf.limit()); } return length; } /** * Reads forward until the signature of the "End of central * directory" record is found. */ private boolean findEocdRecord() throws IOException { int currentByte = -1; boolean skipReadCall = false; while (skipReadCall || (currentByte = readOneByte()) > -1) { skipReadCall = false; if (!isFirstByteOfEocdSig(currentByte)) { continue; } currentByte = readOneByte(); if (currentByte != ZipArchiveOutputStream.EOCD_SIG[1]) { if (currentByte == -1) { break; } skipReadCall = isFirstByteOfEocdSig(currentByte); continue; } currentByte = readOneByte(); if (currentByte != ZipArchiveOutputStream.EOCD_SIG[2]) { if (currentByte == -1) { break; } skipReadCall = isFirstByteOfEocdSig(currentByte); continue; } currentByte = readOneByte(); if (currentByte == -1) { break; } if (currentByte == ZipArchiveOutputStream.EOCD_SIG[3]) { return true; } skipReadCall = isFirstByteOfEocdSig(currentByte); } return false; } /** * Get the number of bytes Inflater has actually processed. * *

for Java < Java7 the getBytes* methods in * Inflater/Deflater seem to return unsigned ints rather than * longs that start over with 0 at 2^32.

* *

The stream knows how many bytes it has read, but not how * many the Inflater actually consumed - it should be between the * total number of bytes read for the entry and the total number * minus the last read operation. Here we just try to make the * value close enough to the bytes we've read by assuming the * number of bytes consumed must be smaller than (or equal to) the * number of bytes read but not smaller by more than 2^32.

*/ private long getBytesInflated() { long inB = inf.getBytesRead(); if (current.bytesReadFromStream >= TWO_EXP_32) { while (inB + TWO_EXP_32 <= current.bytesReadFromStream) { inB += TWO_EXP_32; } } return inB; } /** * @since 1.17 */ @SuppressWarnings("resource") // checkInputStream() does not allocate. @Override public long getCompressedCount() { final int method = current.entry.getMethod(); if (method == ZipArchiveOutputStream.STORED) { return current.bytesRead; } if (method == ZipArchiveOutputStream.DEFLATED) { return getBytesInflated(); } if (method == ZipMethod.UNSHRINKING.getCode() || method == ZipMethod.IMPLODING.getCode() || method == ZipMethod.ENHANCED_DEFLATED.getCode() || method == ZipMethod.BZIP2.getCode()) { return ((InputStreamStatistics) current.checkInputStream()).getCompressedCount(); } return -1; } @Override public ArchiveEntry getNextEntry() throws IOException { return getNextZipEntry(); } public ZipArchiveEntry getNextZipEntry() throws IOException { uncompressedCount = 0; boolean firstEntry = true; if (closed || hitCentralDirectory) { return null; } if (current != null) { closeEntry(); firstEntry = false; } final long currentHeaderOffset = getBytesRead(); try { if (firstEntry) { // split archives have a special signature before the // first local file header - look for it and fail with // the appropriate error message if this is a split // archive. readFirstLocalFileHeader(); } else { readFully(lfhBuf); } } catch (final EOFException e) { //NOSONAR return null; } final ZipLong sig = new ZipLong(lfhBuf); if (!sig.equals(ZipLong.LFH_SIG)) { if (sig.equals(ZipLong.CFH_SIG) || sig.equals(ZipLong.AED_SIG) || isApkSigningBlock(lfhBuf)) { hitCentralDirectory = true; skipRemainderOfArchive(); return null; } throw new ZipException(String.format("Unexpected record signature: 0x%x", sig.getValue())); } int off = WORD; current = new CurrentEntry(); final int versionMadeBy = ZipShort.getValue(lfhBuf, off); off += SHORT; current.entry.setPlatform((versionMadeBy >> ZipFile.BYTE_SHIFT) & ZipFile.NIBLET_MASK); final GeneralPurposeBit gpFlag = GeneralPurposeBit.parse(lfhBuf, off); final boolean hasUTF8Flag = gpFlag.usesUTF8ForNames(); final ZipEncoding entryEncoding = hasUTF8Flag ? ZipEncodingHelper.UTF8_ZIP_ENCODING : zipEncoding; current.hasDataDescriptor = gpFlag.usesDataDescriptor(); current.entry.setGeneralPurposeBit(gpFlag); off += SHORT; current.entry.setMethod(ZipShort.getValue(lfhBuf, off)); off += SHORT; final long time = ZipUtil.dosToJavaTime(ZipLong.getValue(lfhBuf, off)); current.entry.setTime(time); off += WORD; ZipLong size = null, cSize = null; if (!current.hasDataDescriptor) { current.entry.setCrc(ZipLong.getValue(lfhBuf, off)); off += WORD; cSize = new ZipLong(lfhBuf, off); off += WORD; size = new ZipLong(lfhBuf, off); off += WORD; } else { off += 3 * WORD; } final int fileNameLen = ZipShort.getValue(lfhBuf, off); off += SHORT; final int extraLen = ZipShort.getValue(lfhBuf, off); off += SHORT; // NOSONAR - assignment as documentation final byte[] fileName = readRange(fileNameLen); current.entry.setName(entryEncoding.decode(fileName), fileName); if (hasUTF8Flag) { current.entry.setNameSource(ZipArchiveEntry.NameSource.NAME_WITH_EFS_FLAG); } final byte[] extraData = readRange(extraLen); try { current.entry.setExtra(extraData); } catch (final RuntimeException ex) { final ZipException z = new ZipException("Invalid extra data in entry " + current.entry.getName()); z.initCause(ex); throw z; } if (!hasUTF8Flag && useUnicodeExtraFields) { ZipUtil.setNameAndCommentFromExtraFields(current.entry, fileName, null); } processZip64Extra(size, cSize); current.entry.setLocalHeaderOffset(currentHeaderOffset); current.entry.setDataOffset(getBytesRead()); current.entry.setStreamContiguous(true); final ZipMethod m = ZipMethod.getMethodByCode(current.entry.getMethod()); if (current.entry.getCompressedSize() != ArchiveEntry.SIZE_UNKNOWN) { if (ZipUtil.canHandleEntryData(current.entry) && m != ZipMethod.STORED && m != ZipMethod.DEFLATED) { final InputStream bis = new BoundedInputStream(inputStream, current.entry.getCompressedSize()); switch (m) { case UNSHRINKING: current.inputStream = new UnshrinkingInputStream(bis); break; case IMPLODING: try { current.inputStream = new ExplodingInputStream( current.entry.getGeneralPurposeBit().getSlidingDictionarySize(), current.entry.getGeneralPurposeBit().getNumberOfShannonFanoTrees(), bis); } catch (final IllegalArgumentException ex) { throw new IOException("bad IMPLODE data", ex); } break; case BZIP2: current.inputStream = new BZip2CompressorInputStream(bis); break; case ENHANCED_DEFLATED: current.inputStream = new Deflate64CompressorInputStream(bis); break; default: // we should never get here as all supported methods have been covered // will cause an error when read is invoked, don't throw an exception here so people can // skip unsupported entries break; } } } else if (m == ZipMethod.ENHANCED_DEFLATED) { current.inputStream = new Deflate64CompressorInputStream(inputStream); } entriesRead++; return current.entry; } /** * @since 1.17 */ @Override public long getUncompressedCount() { return uncompressedCount; } /** * Checks whether this might be an APK Signing Block. * *

Unfortunately the APK signing block does not start with some kind of signature, it rather ends with one. It * starts with a length, so what we do is parse the suspect length, skip ahead far enough, look for the signature * and if we've found it, return true.

* * @param suspectLocalFileHeader the bytes read from the underlying stream in the expectation that they would hold * the local file header of the next entry. * * @return true if this looks like a APK signing block * * @see https://source.android.com/security/apksigning/v2 */ private boolean isApkSigningBlock(final byte[] suspectLocalFileHeader) throws IOException { // length of block excluding the size field itself final BigInteger len = ZipEightByteInteger.getValue(suspectLocalFileHeader); // LFH has already been read and all but the first eight bytes contain (part of) the APK signing block, // also subtract 16 bytes in order to position us at the magic string BigInteger toSkip = len.add(BigInteger.valueOf(DWORD - suspectLocalFileHeader.length - (long) APK_SIGNING_BLOCK_MAGIC.length)); final byte[] magic = new byte[APK_SIGNING_BLOCK_MAGIC.length]; try { if (toSkip.signum() < 0) { // suspectLocalFileHeader contains the start of suspect magic string final int off = suspectLocalFileHeader.length + toSkip.intValue(); // length was shorter than magic length if (off < DWORD) { return false; } final int bytesInBuffer = Math.abs(toSkip.intValue()); System.arraycopy(suspectLocalFileHeader, off, magic, 0, Math.min(bytesInBuffer, magic.length)); if (bytesInBuffer < magic.length) { readFully(magic, bytesInBuffer); } } else { while (toSkip.compareTo(LONG_MAX) > 0) { realSkip(Long.MAX_VALUE); toSkip = toSkip.add(LONG_MAX.negate()); } realSkip(toSkip.longValue()); readFully(magic); } } catch (final EOFException ex) { //NOSONAR // length was invalid return false; } return Arrays.equals(magic, APK_SIGNING_BLOCK_MAGIC); } private boolean isFirstByteOfEocdSig(final int b) { return b == ZipArchiveOutputStream.EOCD_SIG[0]; } /** * Records whether a Zip64 extra is present and sets the size * information from it if sizes are 0xFFFFFFFF and the entry * doesn't use a data descriptor. */ private void processZip64Extra(final ZipLong size, final ZipLong cSize) throws ZipException { final ZipExtraField extra = current.entry.getExtraField(Zip64ExtendedInformationExtraField.HEADER_ID); if (extra != null && !(extra instanceof Zip64ExtendedInformationExtraField)) { throw new ZipException("archive contains unparseable zip64 extra field"); } final Zip64ExtendedInformationExtraField z64 = (Zip64ExtendedInformationExtraField) extra; current.usesZip64 = z64 != null; if (!current.hasDataDescriptor) { if (z64 != null // same as current.usesZip64 but avoids NPE warning && (ZipLong.ZIP64_MAGIC.equals(cSize) || ZipLong.ZIP64_MAGIC.equals(size)) ) { if (z64.getCompressedSize() == null || z64.getSize() == null) { // avoid NPE if it's a corrupted ZIP archive throw new ZipException("archive contains corrupted zip64 extra field"); } long s = z64.getCompressedSize().getLongValue(); if (s < 0) { throw new ZipException("broken archive, entry with negative compressed size"); } current.entry.setCompressedSize(s); s = z64.getSize().getLongValue(); if (s < 0) { throw new ZipException("broken archive, entry with negative size"); } current.entry.setSize(s); } else if (cSize != null && size != null) { if (cSize.getValue() < 0) { throw new ZipException("broken archive, entry with negative compressed size"); } current.entry.setCompressedSize(cSize.getValue()); if (size.getValue() < 0) { throw new ZipException("broken archive, entry with negative size"); } current.entry.setSize(size.getValue()); } } } private void pushback(final byte[] buf, final int offset, final int length) throws IOException { ((PushbackInputStream) inputStream).unread(buf, offset, length); pushedBackBytes(length); } @Override public int read(final byte[] buffer, final int offset, final int length) throws IOException { if (length == 0) { return 0; } if (closed) { throw new IOException("The stream is closed"); } if (current == null) { return -1; } // avoid int overflow, check null buffer if (offset > buffer.length || length < 0 || offset < 0 || buffer.length - offset < length) { throw new ArrayIndexOutOfBoundsException(); } ZipUtil.checkRequestedFeatures(current.entry); if (!supportsDataDescriptorFor(current.entry)) { throw new UnsupportedZipFeatureException(UnsupportedZipFeatureException.Feature.DATA_DESCRIPTOR, current.entry); } if (!supportsCompressedSizeFor(current.entry)) { throw new UnsupportedZipFeatureException(UnsupportedZipFeatureException.Feature.UNKNOWN_COMPRESSED_SIZE, current.entry); } final int read; if (current.entry.getMethod() == ZipArchiveOutputStream.STORED) { read = readStored(buffer, offset, length); } else if (current.entry.getMethod() == ZipArchiveOutputStream.DEFLATED) { read = readDeflated(buffer, offset, length); } else if (current.entry.getMethod() == ZipMethod.UNSHRINKING.getCode() || current.entry.getMethod() == ZipMethod.IMPLODING.getCode() || current.entry.getMethod() == ZipMethod.ENHANCED_DEFLATED.getCode() || current.entry.getMethod() == ZipMethod.BZIP2.getCode()) { read = current.inputStream.read(buffer, offset, length); } else { throw new UnsupportedZipFeatureException(ZipMethod.getMethodByCode(current.entry.getMethod()), current.entry); } if (read >= 0) { current.crc.update(buffer, offset, read); uncompressedCount += read; } return read; } private void readDataDescriptor() throws IOException { readFully(wordBuf); ZipLong val = new ZipLong(wordBuf); if (ZipLong.DD_SIG.equals(val)) { // data descriptor with signature, skip sig readFully(wordBuf); val = new ZipLong(wordBuf); } current.entry.setCrc(val.getValue()); // if there is a ZIP64 extra field, sizes are eight bytes // each, otherwise four bytes each. Unfortunately some // implementations - namely Java7 - use eight bytes without // using a ZIP64 extra field - // https://bugs.sun.com/bugdatabase/view_bug.do?bug_id=7073588 // just read 16 bytes and check whether bytes nine to twelve // look like one of the signatures of what could follow a data // descriptor (ignoring archive decryption headers for now). // If so, push back eight bytes and assume sizes are four // bytes, otherwise sizes are eight bytes each. readFully(twoDwordBuf); final ZipLong potentialSig = new ZipLong(twoDwordBuf, DWORD); if (potentialSig.equals(ZipLong.CFH_SIG) || potentialSig.equals(ZipLong.LFH_SIG)) { pushback(twoDwordBuf, DWORD, DWORD); long size = ZipLong.getValue(twoDwordBuf); if (size < 0) { throw new ZipException("broken archive, entry with negative compressed size"); } current.entry.setCompressedSize(size); size = ZipLong.getValue(twoDwordBuf, WORD); if (size < 0) { throw new ZipException("broken archive, entry with negative size"); } current.entry.setSize(size); } else { long size = ZipEightByteInteger.getLongValue(twoDwordBuf); if (size < 0) { throw new ZipException("broken archive, entry with negative compressed size"); } current.entry.setCompressedSize(size); size = ZipEightByteInteger.getLongValue(twoDwordBuf, DWORD); if (size < 0) { throw new ZipException("broken archive, entry with negative size"); } current.entry.setSize(size); } } /** * Implementation of read for DEFLATED entries. */ private int readDeflated(final byte[] buffer, final int offset, final int length) throws IOException { final int read = readFromInflater(buffer, offset, length); if (read <= 0) { if (inf.finished()) { return -1; } if (inf.needsDictionary()) { throw new ZipException("This archive needs a preset dictionary" + " which is not supported by Commons" + " Compress."); } if (read == -1) { throw new IOException("Truncated ZIP file"); } } return read; } /** * Fills the given array with the first local file header and * deals with splitting/spanning markers that may prefix the first * LFH. */ private void readFirstLocalFileHeader() throws IOException { readFully(lfhBuf); final ZipLong sig = new ZipLong(lfhBuf); if (!skipSplitSig && sig.equals(ZipLong.DD_SIG)) { throw new UnsupportedZipFeatureException(UnsupportedZipFeatureException.Feature.SPLITTING); } // the split ZIP signature(08074B50) should only be skipped when the skipSplitSig is set if (sig.equals(ZipLong.SINGLE_SEGMENT_SPLIT_MARKER) || sig.equals(ZipLong.DD_SIG)) { // Just skip over the marker. final byte[] missedLfhBytes = new byte[4]; readFully(missedLfhBytes); System.arraycopy(lfhBuf, 4, lfhBuf, 0, LFH_LEN - 4); System.arraycopy(missedLfhBytes, 0, lfhBuf, LFH_LEN - 4, 4); } } /** * Potentially reads more bytes to fill the inflater's buffer and * reads from it. */ private int readFromInflater(final byte[] buffer, final int offset, final int length) throws IOException { int read = 0; do { if (inf.needsInput()) { final int l = fill(); if (l > 0) { current.bytesReadFromStream += buf.limit(); } else if (l == -1) { return -1; } else { break; } } try { read = inf.inflate(buffer, offset, length); } catch (final DataFormatException e) { throw (IOException) new ZipException(e.getMessage()).initCause(e); } } while (read == 0 && inf.needsInput()); return read; } private void readFully(final byte[] b) throws IOException { readFully(b, 0); } // End of Central Directory Record // end of central dir signature WORD // number of this disk SHORT // number of the disk with the // start of the central directory SHORT // total number of entries in the // central directory on this disk SHORT // total number of entries in // the central directory SHORT // size of the central directory WORD // offset of start of central // directory with respect to // the starting disk number WORD // .ZIP file comment length SHORT // .ZIP file comment up to 64KB // private void readFully(final byte[] b, final int off) throws IOException { final int len = b.length - off; final int count = IOUtils.readFully(inputStream, b, off, len); count(count); if (count < len) { throw new EOFException(); } } /** * Reads bytes by reading from the underlying stream rather than * the (potentially inflating) archive stream - which {@link #read} would do. * * Also updates bytes-read counter. */ private int readOneByte() throws IOException { final int b = inputStream.read(); if (b != -1) { count(1); } return b; } private byte[] readRange(final int len) throws IOException { final byte[] ret = IOUtils.readRange(inputStream, len); count(ret.length); if (ret.length < len) { throw new EOFException(); } return ret; } /** * Implementation of read for STORED entries. */ private int readStored(final byte[] buffer, final int offset, final int length) throws IOException { if (current.hasDataDescriptor) { if (lastStoredEntry == null) { readStoredEntry(); } return lastStoredEntry.read(buffer, offset, length); } final long csize = current.entry.getSize(); if (current.bytesRead >= csize) { return -1; } if (buf.position() >= buf.limit()) { buf.position(0); final int l = inputStream.read(buf.array()); if (l == -1) { buf.limit(0); throw new IOException("Truncated ZIP file"); } buf.limit(l); count(l); current.bytesReadFromStream += l; } int toRead = Math.min(buf.remaining(), length); if ((csize - current.bytesRead) < toRead) { // if it is smaller than toRead then it fits into an int toRead = (int) (csize - current.bytesRead); } buf.get(buffer, offset, toRead); current.bytesRead += toRead; return toRead; } /** * Caches a stored entry that uses the data descriptor. * *
    *
  • Reads a stored entry until the signature of a local file * header, central directory header or data descriptor has been * found.
  • *
  • Stores all entry data in lastStoredEntry.

    *
  • Rewinds the stream to position at the data * descriptor.
  • *
  • reads the data descriptor
  • *
* *

After calling this method the entry should know its size, * the entry's data is cached and the stream is positioned at the * next local file or central directory header.

*/ private void readStoredEntry() throws IOException { final ByteArrayOutputStream bos = new ByteArrayOutputStream(); int off = 0; boolean done = false; // length of DD without signature final int ddLen = current.usesZip64 ? WORD + 2 * DWORD : 3 * WORD; while (!done) { final int r = inputStream.read(buf.array(), off, ZipArchiveOutputStream.BUFFER_SIZE - off); if (r <= 0) { // read the whole archive without ever finding a // central directory throw new IOException("Truncated ZIP file"); } if (r + off < 4) { // buffer too small to check for a signature, loop off += r; continue; } done = bufferContainsSignature(bos, off, r, ddLen); if (!done) { off = cacheBytesRead(bos, off, r, ddLen); } } if (current.entry.getCompressedSize() != current.entry.getSize()) { throw new ZipException("compressed and uncompressed size don't match" + USE_ZIPFILE_INSTEAD_OF_STREAM_DISCLAIMER); } final byte[] b = bos.toByteArray(); if (b.length != current.entry.getSize()) { throw new ZipException("actual and claimed size don't match" + USE_ZIPFILE_INSTEAD_OF_STREAM_DISCLAIMER); } lastStoredEntry = new ByteArrayInputStream(b); } /** * Skips bytes by reading from the underlying stream rather than * the (potentially inflating) archive stream - which {@link * #skip} would do. * * Also updates bytes-read counter. */ private void realSkip(final long value) throws IOException { if (value >= 0) { long skipped = 0; while (skipped < value) { final long rem = value - skipped; final int x = inputStream.read(skipBuf, 0, (int) (skipBuf.length > rem ? rem : skipBuf.length)); if (x == -1) { return; } count(x); skipped += x; } return; } throw new IllegalArgumentException(); } /** * Skips over and discards value bytes of data from this input * stream. * *

This implementation may end up skipping over some smaller * number of bytes, possibly 0, if and only if it reaches the end * of the underlying stream.

* *

The actual number of bytes skipped is returned.

* * @param value the number of bytes to be skipped. * @return the actual number of bytes skipped. * @throws IOException - if an I/O error occurs. * @throws IllegalArgumentException - if value is negative. */ @Override public long skip(final long value) throws IOException { if (value >= 0) { long skipped = 0; while (skipped < value) { final long rem = value - skipped; final int x = read(skipBuf, 0, (int) (skipBuf.length > rem ? rem : skipBuf.length)); if (x == -1) { return skipped; } skipped += x; } return skipped; } throw new IllegalArgumentException(); } /** * Reads the stream until it find the "End of central directory * record" and consumes it as well. */ private void skipRemainderOfArchive() throws IOException { // skip over central directory. One LFH has been read too much // already. The calculation discounts file names and extra // data so it will be too short. if (entriesRead > 0) { realSkip((long) entriesRead * CFH_LEN - LFH_LEN); final boolean foundEocd = findEocdRecord(); if (foundEocd) { realSkip((long) ZipFile.MIN_EOCD_SIZE - WORD /* signature */ - SHORT /* comment len */); readFully(shortBuf); // file comment final int commentLen = ZipShort.getValue(shortBuf); if (commentLen >= 0) { realSkip(commentLen); return; } } } throw new IOException("Truncated ZIP file"); } /** * Whether the compressed size for the entry is either known or * not required by the compression method being used. */ private boolean supportsCompressedSizeFor(final ZipArchiveEntry entry) { return entry.getCompressedSize() != ArchiveEntry.SIZE_UNKNOWN || entry.getMethod() == ZipEntry.DEFLATED || entry.getMethod() == ZipMethod.ENHANCED_DEFLATED.getCode() || (entry.getGeneralPurposeBit().usesDataDescriptor() && allowStoredEntriesWithDataDescriptor && entry.getMethod() == ZipEntry.STORED); } /** * Whether this entry requires a data descriptor this library can work with. * * @return true if allowStoredEntriesWithDataDescriptor is true, * the entry doesn't require any data descriptor or the method is * DEFLATED or ENHANCED_DEFLATED. */ private boolean supportsDataDescriptorFor(final ZipArchiveEntry entry) { return !entry.getGeneralPurposeBit().usesDataDescriptor() || (allowStoredEntriesWithDataDescriptor && entry.getMethod() == ZipEntry.STORED) || entry.getMethod() == ZipEntry.DEFLATED || entry.getMethod() == ZipMethod.ENHANCED_DEFLATED.getCode(); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy