This is an automated email from the ASF dual-hosted git repository.

tallison pushed a commit to branch TIKA-4220-branch2x
in repository https://gitbox.apache.org/repos/asf/tika.git

commit ea5ae249b3a99d48bcf961e11384654de5f7eae3
Author: tallison <talli...@apache.org>
AuthorDate: Sat Mar 23 10:48:35 2024 -0400

    TIKA-4220 on branch_2x -- temporary commons-compress workaround
---
 .../org/apache/tika/parser/pkg/PackageParser.java  |  27 +-
 .../detect/zip/DefaultZipContainerDetector.java    |   5 +-
 .../tika/detect/zip/TikaArchiveStreamFactory.java  | 641 +++++++++++++++++++++
 3 files changed, 663 insertions(+), 10 deletions(-)

diff --git 
a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
 
b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
index 86bfbaa88..f45fa23bd 100644
--- 
a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
+++ 
b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
@@ -61,6 +61,7 @@ import org.xml.sax.helpers.AttributesImpl;
 
 import org.apache.tika.config.Field;
 import org.apache.tika.detect.EncodingDetector;
+import org.apache.tika.detect.zip.TikaArchiveStreamFactory;
 import org.apache.tika.exception.EncryptedDocumentException;
 import org.apache.tika.exception.TikaException;
 import org.apache.tika.extractor.EmbeddedDocumentExtractor;
@@ -256,14 +257,24 @@ public class PackageParser extends 
AbstractEncodingDetectorParser {
         String encoding = null;
         try {
             ArchiveStreamFactory factory =
-                    context.get(ArchiveStreamFactory.class, new 
ArchiveStreamFactory());
-            encoding = factory.getEntryEncoding();
-            // At the end we want to close the archive stream to release
-            // any associated resources, but the underlying document stream
-            // should not be closed
-
-            ais = factory.createArchiveInputStream(new 
CloseShieldInputStream(stream));
-
+                    context.get(ArchiveStreamFactory.class);
+            //TODO -- fix this when we next upgrade commons-compress
+            //TODO -- we've probably already detected the stream by here. We 
should
+            //rely on that detection and not re-detect.
+            if (factory != null) {
+                encoding = factory.getEntryEncoding();
+                // At the end we want to close the archive stream to release
+                // any associated resources, but the underlying document stream
+                // should not be closed
+                ais = factory.createArchiveInputStream(new 
CloseShieldInputStream(stream));
+            } else {
+                TikaArchiveStreamFactory tikaFactory = new 
TikaArchiveStreamFactory();
+                encoding = tikaFactory.getEntryEncoding();
+                // At the end we want to close the archive stream to release
+                // any associated resources, but the underlying document stream
+                // should not be closed
+                ais = tikaFactory.createArchiveInputStream(new 
CloseShieldInputStream(stream));
+            }
         } catch (StreamingNotSupportedException sne) {
             // Most archive formats work on streams, but a few need files
             if (sne.getFormat().equals(ArchiveStreamFactory.SEVEN_Z)) {
diff --git 
a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
 
b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
index d2f28e0a0..9842fc0af 100644
--- 
a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
+++ 
b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
@@ -25,7 +25,6 @@ import java.util.ArrayList;
 import java.util.List;
 
 import org.apache.commons.compress.archivers.ArchiveException;
-import org.apache.commons.compress.archivers.ArchiveStreamFactory;
 import 
org.apache.commons.compress.archivers.zip.UnsupportedZipFeatureException;
 import org.apache.commons.compress.archivers.zip.ZipArchiveEntry;
 import org.apache.commons.compress.archivers.zip.ZipArchiveInputStream;
@@ -121,7 +120,9 @@ public class DefaultZipContainerDetector implements 
Detector {
             return TIFF;
         }
         try {
-            String name = ArchiveStreamFactory.detect(new 
UnsynchronizedByteArrayInputStream(prefix, 0, length));
+            String name =
+                    TikaArchiveStreamFactory.detect(new 
UnsynchronizedByteArrayInputStream(prefix,
+                            0, length));
             return PackageConstants.getMediaType(name);
         } catch (ArchiveException e) {
             return MediaType.OCTET_STREAM;
diff --git 
a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/TikaArchiveStreamFactory.java
 
b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/TikaArchiveStreamFactory.java
new file mode 100644
index 000000000..dc2cb908a
--- /dev/null
+++ 
b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/TikaArchiveStreamFactory.java
@@ -0,0 +1,641 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.tika.detect.zip;
+
+import java.io.ByteArrayInputStream;
+import java.io.IOException;
+import java.io.InputStream;
+import java.io.OutputStream;
+import java.security.AccessController;
+import java.security.PrivilegedAction;
+import java.util.Collections;
+import java.util.Locale;
+import java.util.ServiceLoader;
+import java.util.Set;
+import java.util.SortedMap;
+import java.util.TreeMap;
+
+import org.apache.commons.compress.archivers.ArchiveEntry;
+import org.apache.commons.compress.archivers.ArchiveException;
+import org.apache.commons.compress.archivers.ArchiveInputStream;
+import org.apache.commons.compress.archivers.ArchiveOutputStream;
+import org.apache.commons.compress.archivers.ArchiveStreamProvider;
+import org.apache.commons.compress.archivers.StreamingNotSupportedException;
+import org.apache.commons.compress.archivers.ar.ArArchiveInputStream;
+import org.apache.commons.compress.archivers.ar.ArArchiveOutputStream;
+import org.apache.commons.compress.archivers.arj.ArjArchiveInputStream;
+import org.apache.commons.compress.archivers.cpio.CpioArchiveInputStream;
+import org.apache.commons.compress.archivers.cpio.CpioArchiveOutputStream;
+import org.apache.commons.compress.archivers.dump.DumpArchiveInputStream;
+import org.apache.commons.compress.archivers.jar.JarArchiveInputStream;
+import org.apache.commons.compress.archivers.jar.JarArchiveOutputStream;
+import org.apache.commons.compress.archivers.sevenz.SevenZFile;
+import org.apache.commons.compress.archivers.tar.TarArchiveEntry;
+import org.apache.commons.compress.archivers.tar.TarArchiveInputStream;
+import org.apache.commons.compress.archivers.tar.TarArchiveOutputStream;
+import org.apache.commons.compress.archivers.zip.ZipArchiveInputStream;
+import org.apache.commons.compress.archivers.zip.ZipArchiveOutputStream;
+import org.apache.commons.compress.utils.IOUtils;
+import org.apache.commons.compress.utils.Sets;
+
+/**
+ * Factory to create Archive[In|Out]putStreams from names or the first bytes 
of the InputStream.
+ * In order to add other implementations, you should extend
+ * ArchiveStreamFactory and override the appropriate methods (and call their 
implementation from
+ * super of course).
+ * <p>
+ * Compressing a ZIP-File:
+ *
+ * <pre>
+ * final OutputStream out = Files.newOutputStream(output.toPath());
+ * ArchiveOutputStream os = new 
ArchiveStreamFactory().createArchiveOutputStream(ArchiveStreamFactory.ZIP,
+ * out);
+ *
+ * os.putArchiveEntry(new ZipArchiveEntry("testdata/test1.xml"));
+ * IOUtils.copy(Files.newInputStream(file1.toPath()), os);
+ * os.closeArchiveEntry();
+ *
+ * os.putArchiveEntry(new ZipArchiveEntry("testdata/test2.xml"));
+ * IOUtils.copy(Files.newInputStream(file2.toPath()), os);
+ * os.closeArchiveEntry();
+ * os.close();
+ * </pre>
+ * <p>
+ * Decompressing a ZIP-File:
+ *
+ * <pre>
+ * final InputStream is = Files.newInputStream(input.toPath());
+ * ArchiveInputStream in = new 
ArchiveStreamFactory().createArchiveInputStream(ArchiveStreamFactory.ZIP,
+ * is);
+ * ZipArchiveEntry entry = (ZipArchiveEntry) in.getNextEntry();
+ * OutputStream out = 
Files.newOutputStream(dir.toPath().resolve(entry.getName()));
+ * IOUtils.copy(in, out);
+ * out.close();
+ * in.close();
+ * </pre>
+ *
+ * @Immutable provided that the deprecated method setEntryEncoding is not used.
+ * @ThreadSafe even if the deprecated method setEntryEncoding is used
+ */
+public class TikaArchiveStreamFactory implements ArchiveStreamProvider {
+
+    private static final int TAR_HEADER_SIZE = 512;
+
+    private static final int TAR_TEST_ENTRY_COUNT = 10;
+
+    private static final int DUMP_SIGNATURE_SIZE = 32;
+
+    private static final int SIGNATURE_SIZE = 12;
+
+    /**
+     * The singleton instance using the platform default encoding.
+     *
+     * @since 1.21
+     */
+    public static final TikaArchiveStreamFactory DEFAULT = new 
TikaArchiveStreamFactory();
+
+    /**
+     * Constant (value {@value}) used to identify the APK archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String APK = "apk";
+
+    /**
+     * Constant (value {@value}) used to identify the XAPK archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String XAPK = "xapk";
+
+    /**
+     * Constant (value {@value}) used to identify the APKS archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String APKS = "apks";
+
+    /**
+     * Constant (value {@value}) used to identify the APKM archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String APKM = "apkm";
+
+    /**
+     * Constant (value {@value}) used to identify the AR archive format.
+     *
+     * @since 1.1
+     */
+    public static final String AR = "ar";
+
+    /**
+     * Constant (value {@value}) used to identify the ARJ archive format. Not 
supported as an
+     * output stream type.
+     *
+     * @since 1.6
+     */
+    public static final String ARJ = "arj";
+
+    /**
+     * Constant (value {@value}) used to identify the CPIO archive format.
+     *
+     * @since 1.1
+     */
+    public static final String CPIO = "cpio";
+
+    /**
+     * Constant (value {@value}) used to identify the Unix DUMP archive 
format. Not supported as
+     * an output stream type.
+     *
+     * @since 1.3
+     */
+    public static final String DUMP = "dump";
+
+    /**
+     * Constant (value {@value}) used to identify the JAR archive format.
+     *
+     * @since 1.1
+     */
+    public static final String JAR = "jar";
+
+    /**
+     * Constant used to identify the TAR archive format.
+     *
+     * @since 1.1
+     */
+    public static final String TAR = "tar";
+
+    /**
+     * Constant (value {@value}) used to identify the ZIP archive format.
+     *
+     * @since 1.1
+     */
+    public static final String ZIP = "zip";
+
+    /**
+     * Constant (value {@value}) used to identify the 7z archive format.
+     *
+     * @since 1.8
+     */
+    public static final String SEVEN_Z = "7z";
+
+    private static Iterable<ArchiveStreamProvider> 
archiveStreamProviderIterable() {
+        return ServiceLoader.load(ArchiveStreamProvider.class, 
ClassLoader.getSystemClassLoader());
+    }
+
+    /**
+     * Try to determine the type of Archiver
+     *
+     * @param in input stream
+     * @return type of archiver if found
+     * @throws ArchiveException if an archiver cannot be detected in the stream
+     * @since 1.14
+     */
+    public static String detect(final InputStream in) throws ArchiveException {
+        if (in == null) {
+            throw new IllegalArgumentException("Stream must not be null.");
+        }
+
+        if (!in.markSupported()) {
+            throw new IllegalArgumentException("Mark is not supported.");
+        }
+
+        final byte[] signature = new byte[SIGNATURE_SIZE];
+        in.mark(signature.length);
+        int signatureLength = -1;
+        try {
+            signatureLength = IOUtils.readFully(in, signature);
+            in.reset();
+        } catch (final IOException e) {
+            throw new ArchiveException("IOException while reading signature.", 
e);
+        }
+
+        // For now JAR files are detected as ZIP files.
+        if (ZipArchiveInputStream.matches(signature, signatureLength)) {
+            return ZIP;
+        }
+        // For now JAR files are detected as ZIP files.
+        if (JarArchiveInputStream.matches(signature, signatureLength)) {
+            return JAR;
+        }
+        if (ArArchiveInputStream.matches(signature, signatureLength)) {
+            return AR;
+        }
+        if (CpioArchiveInputStream.matches(signature, signatureLength)) {
+            return CPIO;
+        }
+        if (ArjArchiveInputStream.matches(signature, signatureLength)) {
+            return ARJ;
+        }
+        if (SevenZFile.matches(signature, signatureLength)) {
+            return SEVEN_Z;
+        }
+
+        // Dump needs a bigger buffer to check the signature;
+        final byte[] dumpsig = new byte[DUMP_SIGNATURE_SIZE];
+        in.mark(dumpsig.length);
+        try {
+            signatureLength = IOUtils.readFully(in, dumpsig);
+            in.reset();
+        } catch (final IOException e) {
+            throw new ArchiveException("IOException while reading dump 
signature", e);
+        }
+        if (DumpArchiveInputStream.matches(dumpsig, signatureLength)) {
+            return DUMP;
+        }
+
+        // Tar needs an even bigger buffer to check the signature; read the 
first block
+        final byte[] tarHeader = new byte[TAR_HEADER_SIZE];
+        in.mark(tarHeader.length);
+        try {
+            signatureLength = IOUtils.readFully(in, tarHeader);
+            in.reset();
+        } catch (final IOException e) {
+            throw new ArchiveException("IOException while reading tar 
signature", e);
+        }
+        if (TarArchiveInputStream.matches(tarHeader, signatureLength)) {
+            return TAR;
+        }
+
+        // COMPRESS-117
+        if (signatureLength >= TAR_HEADER_SIZE) {
+            try (TarArchiveInputStream inputStream = new TarArchiveInputStream(
+                    new ByteArrayInputStream(tarHeader))) {
+                // COMPRESS-191 - verify the header checksum
+                // COMPRESS-644 - do not allow zero byte file entries
+                TarArchiveEntry entry = inputStream.getNextEntry();
+                // try to find the first non-directory entry within the first 
10 entries.
+                int count = 0;
+                while (entry != null && entry.isDirectory() && 
entry.isCheckSumOK() &&
+                        count++ < TAR_TEST_ENTRY_COUNT) {
+                    entry = inputStream.getNextEntry();
+                }
+                if (entry != null && entry.isCheckSumOK() && 
!entry.isDirectory() &&
+                        entry.getSize() > 0 || count > 0) {
+                    return TAR;
+                }
+            } catch (final Exception e) { // NOPMD NOSONAR
+                // can generate IllegalArgumentException as well as 
IOException auto-detection,
+                // simply not a TAR ignored
+            }
+        }
+        throw new ArchiveException("No Archiver found for the stream 
signature");
+    }
+
+    /**
+     * Constructs a new sorted map from input stream provider names to 
provider objects.
+     *
+     * <p>
+     * The map returned by this method will have one entry for each provider 
for which support is
+     * available in the current Java virtual machine. If two or more
+     * supported provider have the same name then the resulting map will 
contain just one of them;
+     * which one it will contain is not specified.
+     * </p>
+     *
+     * <p>
+     * The invocation of this method, and the subsequent use of the resulting 
map, may cause
+     * time-consuming disk or network I/O operations to occur. This method
+     * is provided for applications that need to enumerate all of the 
available providers, for
+     * example to allow user provider selection.
+     * </p>
+     *
+     * <p>
+     * This method may return different results at different times if new 
providers are dynamically
+     * made available to the current Java virtual machine.
+     * </p>
+     *
+     * @return An immutable, map from names to provider objects
+     * @since 1.13
+     */
+    public static SortedMap<String, ArchiveStreamProvider> 
findAvailableArchiveInputStreamProviders() {
+        return AccessController.doPrivileged(
+                (PrivilegedAction<SortedMap<String, ArchiveStreamProvider>>) 
() -> {
+                    final TreeMap<String, ArchiveStreamProvider> map = new 
TreeMap<>();
+                    putAll(DEFAULT.getInputStreamArchiveNames(), DEFAULT, map);
+                    archiveStreamProviderIterable().forEach(
+                            provider -> 
putAll(provider.getInputStreamArchiveNames(), provider,
+                                    map));
+                    return map;
+                });
+    }
+
+    /**
+     * Constructs a new sorted map from output stream provider names to 
provider objects.
+     *
+     * <p>
+     * The map returned by this method will have one entry for each provider 
for which support is
+     * available in the current Java virtual machine. If two or more
+     * supported provider have the same name then the resulting map will 
contain just one of them;
+     * which one it will contain is not specified.
+     * </p>
+     *
+     * <p>
+     * The invocation of this method, and the subsequent use of the resulting 
map, may cause
+     * time-consuming disk or network I/O operations to occur. This method
+     * is provided for applications that need to enumerate all of the 
available providers, for
+     * example to allow user provider selection.
+     * </p>
+     *
+     * <p>
+     * This method may return different results at different times if new 
providers are dynamically
+     * made available to the current Java virtual machine.
+     * </p>
+     *
+     * @return An immutable, map from names to provider objects
+     * @since 1.13
+     */
+    public static SortedMap<String, ArchiveStreamProvider> 
findAvailableArchiveOutputStreamProviders() {
+        return AccessController.doPrivileged(
+                (PrivilegedAction<SortedMap<String, ArchiveStreamProvider>>) 
() -> {
+                    final TreeMap<String, ArchiveStreamProvider> map = new 
TreeMap<>();
+                    putAll(DEFAULT.getOutputStreamArchiveNames(), DEFAULT, 
map);
+                    archiveStreamProviderIterable().forEach(
+                            provider -> 
putAll(provider.getOutputStreamArchiveNames(), provider,
+                                    map));
+                    return map;
+                });
+    }
+
+    static void putAll(final Set<String> names, final ArchiveStreamProvider 
provider,
+                       final TreeMap<String, ArchiveStreamProvider> map) {
+        names.forEach(name -> map.put(toKey(name), provider));
+    }
+
+    private static String toKey(final String name) {
+        return name.toUpperCase(Locale.ROOT);
+    }
+
+    /**
+     * Entry encoding, null for the default.
+     */
+    private volatile String entryEncoding;
+
+    private SortedMap<String, ArchiveStreamProvider> 
archiveInputStreamProviders;
+
+    private SortedMap<String, ArchiveStreamProvider> 
archiveOutputStreamProviders;
+
+    /**
+     * Constructs an instance using the platform default encoding.
+     */
+    public TikaArchiveStreamFactory() {
+        this(null);
+    }
+
+    /**
+     * Constructs an instance using the specified encoding.
+     *
+     * @param encoding the encoding to be used.
+     * @since 1.10
+     */
+    public TikaArchiveStreamFactory(final String encoding) {
+        this.entryEncoding = encoding;
+    }
+
+    /**
+     * Creates an archive input stream from an input stream, autodetecting the 
archive type from the first
+     * few bytes of the stream. The InputStream must support
+     * marks, like BufferedInputStream.
+     *
+     * @param <I> The {@link ArchiveInputStream} type.
+     * @param in  the input stream
+     * @return the archive input stream
+     * @throws ArchiveException               if the archiver name is not known
+     * @throws StreamingNotSupportedException if the format cannot be read 
from a stream
+     * @throws IllegalArgumentException       if the stream is null or does 
not support mark
+     */
+    public <I extends ArchiveInputStream<? extends ArchiveEntry>> I 
createArchiveInputStream(
+            final InputStream in) throws ArchiveException {
+        return createArchiveInputStream(detect(in), in);
+    }
+
+    /**
+     * Creates an archive input stream from an archiver name and an input 
stream.
+     *
+     * @param <I>          The {@link ArchiveInputStream} type.
+     * @param archiverName the archive name, i.e. {@value #AR}, {@value #ARJ}, 
{@value #ZIP}, {@value #TAR},
+     * {@value #JAR}, {@value #CPIO}, {@value #DUMP} or
+     *                     {@value #SEVEN_Z}
+     * @param in           the input stream
+     * @return the archive input stream
+     * @throws ArchiveException               if the archiver name is not known
+     * @throws StreamingNotSupportedException if the format cannot be read 
from a stream
+     * @throws IllegalArgumentException       if the archiver name or stream 
is null
+     */
+    public <I extends ArchiveInputStream<? extends ArchiveEntry>> I 
createArchiveInputStream(
+            final String archiverName, final InputStream in) throws 
ArchiveException {
+        return createArchiveInputStream(archiverName, in, entryEncoding);
+    }
+
+    @SuppressWarnings("unchecked")
+    @Override
+    public <I extends ArchiveInputStream<? extends ArchiveEntry>> I 
createArchiveInputStream(
+            final String archiverName, final InputStream in, final String 
actualEncoding)
+            throws ArchiveException {
+
+        if (archiverName == null) {
+            throw new IllegalArgumentException("Archiver name must not be 
null.");
+        }
+
+        if (in == null) {
+            throw new IllegalArgumentException("InputStream must not be 
null.");
+        }
+
+        if (AR.equalsIgnoreCase(archiverName)) {
+            return (I) new ArArchiveInputStream(in);
+        }
+        if (ARJ.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new ArjArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new ArjArchiveInputStream(in);
+        }
+        if (ZIP.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new ZipArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new ZipArchiveInputStream(in);
+        }
+        if (TAR.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new TarArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new TarArchiveInputStream(in);
+        }
+        if (JAR.equalsIgnoreCase(archiverName) || 
APK.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new JarArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new JarArchiveInputStream(in);
+        }
+        if (CPIO.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new CpioArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new CpioArchiveInputStream(in);
+        }
+        if (DUMP.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new DumpArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new DumpArchiveInputStream(in);
+        }
+        if (SEVEN_Z.equalsIgnoreCase(archiverName)) {
+            throw new StreamingNotSupportedException(SEVEN_Z);
+        }
+
+        final ArchiveStreamProvider archiveStreamProvider =
+                getArchiveInputStreamProviders().get(toKey(archiverName));
+        if (archiveStreamProvider != null) {
+            return 
archiveStreamProvider.createArchiveInputStream(archiverName, in, 
actualEncoding);
+        }
+
+        throw new ArchiveException("Archiver: " + archiverName + " not 
found.");
+    }
+
+    /**
+     * Creates an archive output stream from an archiver name and an output 
stream.
+     *
+     * @param <O>          The {@link ArchiveOutputStream} type.
+     * @param archiverName the archive name, i.e. {@value #AR}, {@value #ZIP}, 
{@value #TAR},
+     * {@value #JAR} or {@value #CPIO}
+     * @param out          the output stream
+     * @return the archive output stream
+     * @throws ArchiveException               if the archiver name is not known
+     * @throws StreamingNotSupportedException if the format cannot be written 
to a stream
+     * @throws IllegalArgumentException       if the archiver name or stream 
is null
+     */
+    public <O extends ArchiveOutputStream<? extends ArchiveEntry>> O 
createArchiveOutputStream(
+            final String archiverName, final OutputStream out) throws 
ArchiveException {
+        return createArchiveOutputStream(archiverName, out, entryEncoding);
+    }
+
+    @SuppressWarnings("unchecked")
+    @Override
+    public <O extends ArchiveOutputStream<? extends ArchiveEntry>> O 
createArchiveOutputStream(
+            final String archiverName, final OutputStream out, final String 
actualEncoding)
+            throws ArchiveException {
+        if (archiverName == null) {
+            throw new IllegalArgumentException("Archiver name must not be 
null.");
+        }
+        if (out == null) {
+            throw new IllegalArgumentException("OutputStream must not be 
null.");
+        }
+
+        if (AR.equalsIgnoreCase(archiverName)) {
+            return (O) new ArArchiveOutputStream(out);
+        }
+        if (ZIP.equalsIgnoreCase(archiverName)) {
+            final ZipArchiveOutputStream zip = new ZipArchiveOutputStream(out);
+            if (actualEncoding != null) {
+                zip.setEncoding(actualEncoding);
+            }
+            return (O) zip;
+        }
+        if (TAR.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (O) new TarArchiveOutputStream(out, actualEncoding);
+            }
+            return (O) new TarArchiveOutputStream(out);
+        }
+        if (JAR.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (O) new JarArchiveOutputStream(out, actualEncoding);
+            }
+            return (O) new JarArchiveOutputStream(out);
+        }
+        if (CPIO.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (O) new CpioArchiveOutputStream(out, actualEncoding);
+            }
+            return (O) new CpioArchiveOutputStream(out);
+        }
+        if (SEVEN_Z.equalsIgnoreCase(archiverName)) {
+            throw new StreamingNotSupportedException(SEVEN_Z);
+        }
+
+        final ArchiveStreamProvider archiveStreamProvider =
+                getArchiveOutputStreamProviders().get(toKey(archiverName));
+        if (archiveStreamProvider != null) {
+            return 
archiveStreamProvider.createArchiveOutputStream(archiverName, out,
+                    actualEncoding);
+        }
+
+        throw new ArchiveException("Archiver: " + archiverName + " not 
found.");
+    }
+
+    public SortedMap<String, ArchiveStreamProvider> 
getArchiveInputStreamProviders() {
+        if (archiveInputStreamProviders == null) {
+            archiveInputStreamProviders =
+                    
Collections.unmodifiableSortedMap(findAvailableArchiveInputStreamProviders());
+        }
+        return archiveInputStreamProviders;
+    }
+
+    public SortedMap<String, ArchiveStreamProvider> 
getArchiveOutputStreamProviders() {
+        if (archiveOutputStreamProviders == null) {
+            archiveOutputStreamProviders =
+                    
Collections.unmodifiableSortedMap(findAvailableArchiveOutputStreamProviders());
+        }
+        return archiveOutputStreamProviders;
+    }
+
+    /**
+     * Gets the encoding to use for arj, jar, ZIP, dump, cpio and tar files, 
or null for the archiver default.
+     *
+     * @return entry encoding, or null for the archiver default
+     * @since 1.5
+     */
+    public String getEntryEncoding() {
+        return entryEncoding;
+    }
+
+    @Override
+    public Set<String> getInputStreamArchiveNames() {
+        return Sets.newHashSet(AR, ARJ, ZIP, TAR, JAR, CPIO, DUMP, SEVEN_Z);
+    }
+
+    @Override
+    public Set<String> getOutputStreamArchiveNames() {
+        return Sets.newHashSet(AR, ZIP, TAR, JAR, CPIO, SEVEN_Z);
+    }
+
+    /**
+     * Sets the encoding to use for arj, jar, ZIP, dump, cpio and tar files. 
Use null for the archiver default.
+     *
+     * @param entryEncoding the entry encoding, null uses the archiver default.
+     * @since 1.5
+     * @deprecated 1.10 use {@link #ArchiveStreamFactory(String)} to specify 
the encoding
+     */
+    @Deprecated
+    public void setEntryEncoding(final String entryEncoding) {
+        this.entryEncoding = entryEncoding;
+    }
+
+}

Reply via email to