All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.amazonaws.mobileconnectors.s3.transfermanager.TransferManager Maven / Gradle / Ivy

Go to download

The AWS Java Mobile SDK for Amazon S3 module holds the client classes that are used for communicating with Amazon Simple Storage Service

The newest version!
/*
 * Copyright 2010-2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.
 *
 * Licensed under the Apache License, Version 2.0 (the "License").
 * You may not use this file except in compliance with the License.
 * A copy of the License is located at
 *
 *  http://aws.amazon.com/apache2.0
 *
 * or in the "license" file accompanying this file. This file is distributed
 * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
 * express or implied. See the License for the specific language governing
 * permissions and limitations under the License.
 */

package com.amazonaws.mobileconnectors.s3.transfermanager;

import static com.amazonaws.services.s3.internal.ServiceUtils.APPEND_MODE;
import static com.amazonaws.services.s3.internal.ServiceUtils.OVERWRITE_MODE;

import com.amazonaws.AbortedException;
import com.amazonaws.AmazonClientException;
import com.amazonaws.AmazonServiceException;
import com.amazonaws.AmazonWebServiceRequest;
import com.amazonaws.auth.AWSCredentials;
import com.amazonaws.auth.AWSCredentialsProvider;
import com.amazonaws.auth.DefaultAWSCredentialsProviderChain;
import com.amazonaws.event.ProgressEvent;
import com.amazonaws.event.ProgressListener;
import com.amazonaws.event.ProgressListenerChain;
import com.amazonaws.event.ProgressListenerChain.ProgressEventFilter;
import com.amazonaws.mobileconnectors.s3.transfermanager.Transfer.TransferState;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.CopyCallable;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.CopyImpl;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.CopyMonitor;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.DownloadImpl;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.DownloadMonitor;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.MultipleFileDownloadImpl;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.MultipleFileTransferMonitor;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.MultipleFileUploadImpl;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.S3ProgressListener;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.S3ProgressListenerChain;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.TransferManagerUtils;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.TransferProgressUpdatingListener;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.TransferStateChangeListener;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.UploadCallable;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.UploadImpl;
import com.amazonaws.mobileconnectors.s3.transfermanager.internal.UploadMonitor;
import com.amazonaws.services.s3.AmazonS3;
import com.amazonaws.services.s3.AmazonS3Client;
import com.amazonaws.services.s3.AmazonS3EncryptionClient;
import com.amazonaws.services.s3.internal.ServiceUtils;
import com.amazonaws.services.s3.model.AbortMultipartUploadRequest;
import com.amazonaws.services.s3.model.CopyObjectRequest;
import com.amazonaws.services.s3.model.GetObjectMetadataRequest;
import com.amazonaws.services.s3.model.GetObjectRequest;
import com.amazonaws.services.s3.model.ListMultipartUploadsRequest;
import com.amazonaws.services.s3.model.ListObjectsRequest;
import com.amazonaws.services.s3.model.MultipartUpload;
import com.amazonaws.services.s3.model.MultipartUploadListing;
import com.amazonaws.services.s3.model.ObjectListing;
import com.amazonaws.services.s3.model.ObjectMetadata;
import com.amazonaws.services.s3.model.PutObjectRequest;
import com.amazonaws.services.s3.model.S3Object;
import com.amazonaws.services.s3.model.S3ObjectSummary;
import com.amazonaws.services.s3.util.Mimetypes;
import com.amazonaws.util.VersionInfoUtils;

import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;

import java.io.File;
import java.io.InputStream;
import java.util.ArrayList;
import java.util.Date;
import java.util.LinkedList;
import java.util.List;
import java.util.Stack;
import java.util.concurrent.Callable;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Future;
import java.util.concurrent.ScheduledExecutorService;
import java.util.concurrent.ScheduledThreadPoolExecutor;
import java.util.concurrent.ThreadFactory;
import java.util.concurrent.atomic.AtomicInteger;

/**
 * High level utility for managing transfers to Amazon S3.
 * 

* TransferManager provides a simple API for uploading content to * Amazon S3, and makes extensive use of Amazon S3 multipart uploads to achieve * enhanced throughput, performance and reliability. *

* When possible, TransferManager attempts to use multiple threads * to upload multiple parts of a single upload at once. When dealing with large * content sizes and high bandwidth, this can have a significant increase on * throughput. *

* TransferManager is responsible for managing resources such as * connections and threads; share a single instance of * TransferManager whenever possible. TransferManager, * like all the client classes in the AWS SDK for Java, is thread safe. Call * TransferManager.shutdownNow() to release the resources once the * transfer is complete. *

* Using TransferManager to upload options to Amazon S3 is easy: * *

 * DefaultAWSCredentialsProviderChain credentialProviderChain = new DefaultAWSCredentialsProviderChain();
 * TransferManager tx = new TransferManager(
 *         credentialProviderChain.getCredentials());
 * Upload myUpload = tx.upload(myBucket, myFile.getName(), myFile);
 *
 * // You can poll your transfer's status to check its progress
 * if (myUpload.isDone() == false) {
 *     System.out.println("Transfer: " + myUpload.getDescription());
 *     System.out.println("  - State: " + myUpload.getState());
 *     System.out.println("  - Progress: "
 *             + myUpload.getProgress().getBytesTransferred());
 * }
 *
 * // Transfers also allow you to set a <code>ProgressListener</code> to receive
 * // asynchronous notifications about your transfer's progress.
 * myUpload.addProgressListener(myProgressListener);
 *
 * // Or you can block the current thread and wait for your transfer to
 * // to complete. If the transfer fails, this method will throw an
 * // AmazonClientException or AmazonServiceException detailing the reason.
 * myUpload.waitForCompletion();
 *
 * // After the upload is complete, call shutdownNow to release the resources.
 * tx.shutdownNow();
 * 
*

* Transfers can be paused and resumed at a later time. It can also survive JVM * crash, provided the information that is required to resume the transfer is * given as input to the resume operation. For more information on pause and * resume, * * @see Upload#pause() * @see Download#pause() * @see TransferManager#resumeUpload(PersistableUpload) * @see TransferManager#resumeDownload(PersistableDownload) * @deprecated The Transfer Manager is now deprecated in favor of the * {@link com.amazonaws.mobileconnectors.s3.transferutility.TransferUtility} */ @Deprecated public class TransferManager { /** The low level client we use to make the actual calls to Amazon S3. */ private final AmazonS3 s3; /** Configuration for how TransferManager processes requests. */ private TransferManagerConfiguration configuration; /** The thread pool in which transfers are uploaded or downloaded. */ private final ExecutorService threadPool; /** * Thread used for periodicially checking transfers and updating thier * state. */ private final ScheduledExecutorService timedThreadPool = new ScheduledThreadPoolExecutor(1, daemonThreadFactory); private static final Log log = LogFactory.getLog(TransferManager.class); /** * Constructs a new TransferManager and Amazon S3 client using * the credentials from DefaultAWSCredentialsProviderChain *

* TransferManager and client objects may pool connections and * threads. Reuse TransferManager and client objects and share * them throughout applications. *

* TransferManager and all AWS client objects are thread safe. */ public TransferManager() { this(new AmazonS3Client(new DefaultAWSCredentialsProviderChain())); } /** * Constructs a new TransferManager and Amazon S3 client using * the specified AWS security credentials provider. *

* TransferManager and client objects may pool connections and * threads. Reuse TransferManager and client objects and share * them throughout applications. *

* TransferManager and all AWS client objects are thread safe. * * @param credentialsProvider The AWS security credentials provider to use * when making authenticated requests. */ public TransferManager(AWSCredentialsProvider credentialsProvider) { this(new AmazonS3Client(credentialsProvider)); } /** * Constructs a new TransferManager and Amazon S3 client using * the specified AWS security credentials. *

* TransferManager and client objects may pool connections and * threads. Reuse TransferManager and client objects and share * them throughout applications. *

* TransferManager and all AWS client objects are thread safe. * * @param credentials The AWS security credentials to use when making * authenticated requests. */ public TransferManager(AWSCredentials credentials) { this(new AmazonS3Client(credentials)); } /** * Constructs a new TransferManager, specifying the client to * use when making requests to Amazon S3. *

* TransferManager and client objects may pool connections and * threads. Reuse TransferManager and client objects and share * them throughout applications. *

* TransferManager and all AWS client objects are thread safe. *

* * @param s3 The client to use when making requests to Amazon S3. */ public TransferManager(AmazonS3 s3) { this(s3, TransferManagerUtils.createDefaultExecutorService()); } /** * Constructs a new TransferManager specifying the client and * thread pool to use when making requests to Amazon S3. *

* TransferManager and client objects may pool connections and * threads. Reuse TransferManager and client objects and share * them throughout applications. *

* TransferManager and all AWS client objects are thread safe. * * @param s3 The client to use when making requests to Amazon S3. * @param threadPool The thread pool in which to execute requests. */ public TransferManager(AmazonS3 s3, ExecutorService threadPool) { this.s3 = s3; this.threadPool = threadPool; this.configuration = new TransferManagerConfiguration(); } /** * Sets the configuration which specifies how this * TransferManager processes requests. * * @param configuration The new configuration specifying how this * TransferManager processes requests. */ public void setConfiguration(TransferManagerConfiguration configuration) { this.configuration = configuration; } /** * Returns the configuration which specifies how this * TransferManager processes requests. * * @return The configuration settings for this TransferManager. */ public TransferManagerConfiguration getConfiguration() { return configuration; } /** * Returns the underlying Amazon S3 client used to make requests to Amazon * S3. * * @return The underlying Amazon S3 client used to make requests to Amazon * S3. */ public AmazonS3 getAmazonS3Client() { return s3; } /** *

* Schedules a new transfer to upload data to Amazon S3. This method is * non-blocking and returns immediately (i.e. before the upload has * finished). *

*

* When uploading options from a stream, callers must supply the size * of options in the stream through the content length field in the * ObjectMetadata parameter. If no content length is specified * for the input stream, then TransferManager will attempt to buffer all the * stream contents in memory and upload the options as a traditional, single * part upload. Because the entire stream contents must be buffered in * memory, this can be very expensive, and should be avoided whenever * possible. *

*

* Use the returned Upload object to query the progress of the * transfer, add listeners for progress events, and wait for the upload to * complete. *

*

* If resources are available, the upload will begin immediately. Otherwise, * the upload is scheduled and started as soon as resources become * available. *

* * @param bucketName The name of the bucket to upload the new object to. * @param key The key in the specified bucket by which to store the new * object. * @param input The input stream containing the options to upload to Amazon * S3. * @param objectMetadata Additional information about the object being * uploaded, including the size of the options, content type, * additional custom user metadata, etc. * @return A new Upload object to use to check the state of the * upload, listen for progress notifications, and otherwise manage * the upload. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Upload upload(final String bucketName, final String key, final InputStream input, ObjectMetadata objectMetadata) throws AmazonServiceException, AmazonClientException { return upload(new PutObjectRequest(bucketName, key, input, objectMetadata)); } /** * Schedules a new transfer to upload data to Amazon S3. This method is * non-blocking and returns immediately (i.e. before the upload has * finished). *

* The returned Upload object allows you to query the progress of the * transfer, add listeners for progress events, and wait for the upload to * complete. *

* If resources are available, the upload will begin immediately, otherwise * it will be scheduled and started as soon as resources become available. * * @param bucketName The name of the bucket to upload the new object to. * @param key The key in the specified bucket by which to store the new * object. * @param file The file to upload. * @return A new Upload object which can be used to check state of the * upload, listen for progress notifications, and otherwise manage * the upload. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Upload upload(final String bucketName, final String key, final File file) throws AmazonServiceException, AmazonClientException { return upload(new PutObjectRequest(bucketName, key, file)); } /** *

* Schedules a new transfer to upload data to Amazon S3. This method is * non-blocking and returns immediately (i.e. before the upload has * finished). *

*

* Use the returned Upload object to query the progress of the * transfer, add listeners for progress events, and wait for the upload to * complete. *

*

* If resources are available, the upload will begin immediately. Otherwise, * the upload is scheduled and started as soon as resources become * available. *

* * @param putObjectRequest The request containing all the parameters for the * upload. * @return A new Upload object to use to check the state of the * upload, listen for progress notifications, and otherwise manage * the upload. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Upload upload(final PutObjectRequest putObjectRequest) throws AmazonServiceException, AmazonClientException { return doUpload(putObjectRequest, null, null, null); } /** *

* Schedules a new transfer to upload data to Amazon S3. This method is * non-blocking and returns immediately (i.e. before the upload has * finished). *

*

* Use the returned Upload object to query the progress of the * transfer, add listeners for progress events, and wait for the upload to * complete. *

*

* If resources are available, the upload will begin immediately. Otherwise, * the upload is scheduled and started as soon as resources become * available. *

* * @param putObjectRequest The request containing all the parameters for the * upload. * @param progressListener An optional callback listener to receive the * progress of the upload. * @return A new Upload object to use to check the state of the * upload, listen for progress notifications, and otherwise manage * the upload. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Upload upload(final PutObjectRequest putObjectRequest, final S3ProgressListener progressListener) throws AmazonServiceException, AmazonClientException { return doUpload(putObjectRequest, null, progressListener, null); } /** *

* Schedules a new transfer to upload data to Amazon S3. This method is * non-blocking and returns immediately (i.e. before the upload has * finished). *

*

* Use the returned Upload object to query the progress of the * transfer, add listeners for progress events, and wait for the upload to * complete. *

*

* If resources are available, the upload will begin immediately. Otherwise, * the upload is scheduled and started as soon as resources become * available. *

* * @param putObjectRequest The request containing all the parameters for the * upload. * @param stateListener The transfer state change listener to monitor the * upload. * @param progressListener An optional callback listener to receive the * progress of the upload. * @return A new Upload object to use to check the state of the * upload, listen for progress notifications, and otherwise manage * the upload. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ private Upload doUpload(final PutObjectRequest putObjectRequest, final TransferStateChangeListener stateListener, final S3ProgressListener progressListener, final PersistableUpload persistableUpload) throws AmazonServiceException, AmazonClientException { appendSingleObjectUserAgent(putObjectRequest); final String multipartUploadId = persistableUpload != null ? persistableUpload .getMultipartUploadId() : null; if (putObjectRequest.getMetadata() == null) { putObjectRequest.setMetadata(new ObjectMetadata()); } final ObjectMetadata metadata = putObjectRequest.getMetadata(); final File file = TransferManagerUtils.getRequestFile(putObjectRequest); if (file != null) { // Always set the content length, even if it's already set metadata.setContentLength(file.length()); // Only set the content type if it hasn't already been set if (metadata.getContentType() == null) { metadata.setContentType(Mimetypes.getInstance().getMimetype(file)); } } else { if (multipartUploadId != null) { throw new IllegalArgumentException( "Unable to resume the upload. No file specified."); } } final String description = "Uploading to " + putObjectRequest.getBucketName() + "/" + putObjectRequest.getKey(); final TransferProgress transferProgress = new TransferProgress(); transferProgress.setTotalBytesToTransfer(TransferManagerUtils .getContentLength(putObjectRequest)); final S3ProgressListenerChain listenerChain = new S3ProgressListenerChain( new TransferProgressUpdatingListener(transferProgress), putObjectRequest.getGeneralProgressListener(), progressListener); putObjectRequest.setGeneralProgressListener(listenerChain); final UploadImpl upload = new UploadImpl(description, transferProgress, listenerChain, stateListener); final UploadCallable uploadCallable = new UploadCallable(this, threadPool, upload, putObjectRequest, listenerChain, multipartUploadId, transferProgress); final UploadMonitor watcher = new UploadMonitor(this, upload, threadPool, uploadCallable, putObjectRequest, listenerChain); watcher.setTimedThreadPool(timedThreadPool); upload.setMonitor(watcher); return upload; } /** * Schedules a new transfer to download data from Amazon S3 and save it to * the specified file. This method is non-blocking and returns immediately * (i.e. before the data has been fully downloaded). *

* Use the returned Download object to query the progress of the transfer, * add listeners for progress events, and wait for the download to complete. * * @param bucket The name of the bucket containing the object to download. * @param key The key under which the object to download is stored. * @param file The file to download the object's data to. * @return A new Download object to use to check the state of * the download, listen for progress notifications, and otherwise * manage the download. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Download download(String bucket, String key, File file) { return download(new GetObjectRequest(bucket, key), file); } /** * Schedules a new transfer to download data from Amazon S3 and save it to * the specified file. This method is non-blocking and returns immediately * (i.e. before the data has been fully downloaded). *

* Use the returned Download object to query the progress of the transfer, * add listeners for progress events, and wait for the download to complete. * * @param getObjectRequest The request containing all the parameters for the * download. * @param file The file to download the object data to. * @return A new Download object to use to check the state of * the download, listen for progress notifications, and otherwise * manage the download. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Download download(final GetObjectRequest getObjectRequest, final File file) { return doDownload(getObjectRequest, file, null, null, OVERWRITE_MODE); } /** * Schedules a new transfer to download data from Amazon S3 and save it to * the specified file. This method is non-blocking and returns immediately * (i.e. before the data has been fully downloaded). *

* Use the returned Download object to query the progress of the transfer, * add listeners for progress events, and wait for the download to complete. * * @param getObjectRequest The request containing all the parameters for the * download. * @param file The file to download the object data to. * @param progressListener An optional callback listener to get the progress * of the download. * @return A new Download object to use to check the state of * the download, listen for progress notifications, and otherwise * manage the download. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Download download(final GetObjectRequest getObjectRequest, final File file, final S3ProgressListener progressListener) { return doDownload(getObjectRequest, file, null, progressListener, OVERWRITE_MODE); } /** * Same as public interface, but adds a state listener so that callers can * be notified of state changes to the download. * * @see TransferManager#download(GetObjectRequest, File) */ private Download doDownload(final GetObjectRequest getObjectRequest, final File file, final TransferStateChangeListener stateListener, final S3ProgressListener s3progressListener, final boolean resumeExistingDownload) { appendSingleObjectUserAgent(getObjectRequest); final String description = "Downloading from " + getObjectRequest.getBucketName() + "/" + getObjectRequest.getKey(); final TransferProgress transferProgress = new TransferProgress(); // S3 progress listener to capture the persistable transfer when // available final S3ProgressListenerChain listenerChain = new S3ProgressListenerChain( new TransferProgressUpdatingListener(transferProgress), // The // listener // for // updating // transfer // progress getObjectRequest.getGeneralProgressListener(), s3progressListener); // Listeners included in the original // request // The listener chain used by the low-level GetObject request. // This listener chain ignores any COMPLETE event, so that we could // delay firing the signal until the high-level download fully finishes. final ProgressListenerChain listeners = new ProgressListenerChain( new ProgressEventFilter() { @Override public ProgressEvent filter(ProgressEvent progressEvent) { if (progressEvent.getEventCode() == ProgressEvent.COMPLETED_EVENT_CODE) { // Block COMPLETE events from the low-level // GetObject operation, // but we still want to keep the BytesTransferred progressEvent.setEventCode(0); } return progressEvent; } }, listenerChain); getObjectRequest.setGeneralProgressListener(listeners); final GetObjectMetadataRequest getObjectMetadataRequest = new GetObjectMetadataRequest( getObjectRequest.getBucketName(), getObjectRequest.getKey()); if (getObjectRequest.getSSECustomerKey() != null) { getObjectMetadataRequest.setSSECustomerKey(getObjectRequest.getSSECustomerKey()); } final ObjectMetadata objectMetadata = s3.getObjectMetadata(getObjectMetadataRequest); // We still pass the unfiltered listener chain into DownloadImpl final DownloadImpl download = new DownloadImpl(description, transferProgress, listenerChain, null, stateListener, getObjectRequest, file); long startingByte = 0; long lastByte = objectMetadata.getContentLength() - 1; if (getObjectRequest.getRange() != null && getObjectRequest.getRange().length == 2) { startingByte = getObjectRequest.getRange()[0]; lastByte = getObjectRequest.getRange()[1]; } long totalBytesToDownload = lastByte - startingByte + 1; transferProgress.setTotalBytesToTransfer(totalBytesToDownload); if (resumeExistingDownload) { if (file.exists()) { final long numberOfBytesRead = file.length(); startingByte = startingByte + numberOfBytesRead; getObjectRequest.setRange(startingByte, lastByte); transferProgress.updateProgress(Math.min(numberOfBytesRead, totalBytesToDownload)); totalBytesToDownload = lastByte - startingByte + 1; } } if (totalBytesToDownload < 0) { throw new IllegalArgumentException( "Unable to determine the range for download operation."); } final CountDownLatch latch = new CountDownLatch(1); final Future future = submitDownloadTask(getObjectRequest, file, resumeExistingDownload, latch, download); download.setMonitor(new DownloadMonitor(download, future)); latch.countDown(); return download; } private Future submitDownloadTask( final GetObjectRequest getObjectRequest, final File file, final boolean resumeExistingDownload, final CountDownLatch latch, final DownloadImpl download) { final Future future = threadPool.submit(new Callable() { @Override public Object call() throws Exception { try { latch.await(); download.setState(TransferState.InProgress); final S3Object s3Object = ServiceUtils.retryableDownloadS3ObjectToFile(file, new ServiceUtils.RetryableS3DownloadTask() { @Override public S3Object getS3ObjectStream() { final S3Object s3Object = s3.getObject(getObjectRequest); download.setS3Object(s3Object); return s3Object; } @Override public boolean needIntegrityCheck() { // Don't perform the integrity check if the // stream data is wrapped // in a decryption stream, or if we're only // looking at a range of // the data, since otherwise the checksum // won't match up. boolean performIntegrityCheck = true; if (ServiceUtils.skipMd5CheckPerRequest(getObjectRequest) || s3 instanceof AmazonS3EncryptionClient) { performIntegrityCheck = false; } return performIntegrityCheck; } }, resumeExistingDownload); if (s3Object == null) { download.setState(TransferState.Canceled); download.setMonitor(new DownloadMonitor(download, null)); return download; } download.setState(TransferState.Completed); return true; } catch (final Throwable t) { // Downloads aren't allowed to move from canceled to failed if (download.getState() != TransferState.Canceled) { download.setState(TransferState.Failed); } if (t instanceof Exception) { throw (Exception) t; } else { throw (Error) t; } } } }); return future; } /** * Downloads all objects in the virtual directory designated by the * keyPrefix given to the destination directory given. All virtual * subdirectories will be downloaded recursively. * * @param bucketName The bucket containing the virtual directory * @param keyPrefix The key prefix for the virtual directory, or null for * the entire bucket. All subdirectories will be downloaded * recursively. * @param destinationDirectory The directory to place downloaded files. * Subdirectories will be created as necessary. */ public MultipleFileDownload downloadDirectory(String bucketName, String keyPrefix, File destinationDirectory) { if (keyPrefix == null) { keyPrefix = ""; } final List objectSummaries = new LinkedList(); final Stack commonPrefixes = new Stack(); commonPrefixes.add(keyPrefix); long totalSize = 0; // Recurse all virtual subdirectories to get a list of object summaries. // This is a depth-first search. do { final String prefix = commonPrefixes.pop(); ObjectListing listObjectsResponse = null; do { if (listObjectsResponse == null) { final ListObjectsRequest listObjectsRequest = new ListObjectsRequest() .withBucketName(bucketName) .withDelimiter(DEFAULT_DELIMITER).withPrefix(prefix); listObjectsResponse = s3.listObjects(listObjectsRequest); } else { listObjectsResponse = s3.listNextBatchOfObjects(listObjectsResponse); } for (final S3ObjectSummary s : listObjectsResponse.getObjectSummaries()) { // Skip any files that are also virtual directories, since // we can't save both a directory and a file of the same // name. if (!s.getKey().equals(prefix) && !listObjectsResponse.getCommonPrefixes().contains( s.getKey() + DEFAULT_DELIMITER)) { objectSummaries.add(s); totalSize += s.getSize(); } else { log.debug("Skipping download for object " + s.getKey() + " since it is also a virtual directory"); } } commonPrefixes.addAll(listObjectsResponse.getCommonPrefixes()); } while (listObjectsResponse.isTruncated()); } while (!commonPrefixes.isEmpty()); /* This is the hook for adding additional progress listeners */ final ProgressListenerChain additionalListeners = new ProgressListenerChain(); final TransferProgress transferProgress = new TransferProgress(); transferProgress.setTotalBytesToTransfer(totalSize); /* * Bind additional progress listeners to this * MultipleFileTransferProgressUpdatingListener to receive * ByteTransferred events from each single-file download implementation. */ final ProgressListener listener = new MultipleFileTransferProgressUpdatingListener( transferProgress, additionalListeners); final List downloads = new ArrayList(); final String description = "Downloading from " + bucketName + "/" + keyPrefix; final MultipleFileDownloadImpl multipleFileDownload = new MultipleFileDownloadImpl( description, transferProgress, additionalListeners, keyPrefix, bucketName, downloads); multipleFileDownload.setMonitor(new MultipleFileTransferMonitor(multipleFileDownload, downloads)); final CountDownLatch latch = new CountDownLatch(1); final MultipleFileTransferStateChangeListener transferListener = new MultipleFileTransferStateChangeListener( latch, multipleFileDownload); for (final S3ObjectSummary summary : objectSummaries) { // TODO: non-standard delimiters final File f = new File(destinationDirectory, summary.getKey()); final File parentFile = f.getParentFile(); if (!parentFile.exists() && !parentFile.mkdirs()) { throw new RuntimeException("Couldn't create parent directories for " + f.getAbsolutePath()); } // All the single-file downloads share the same // MultipleFileTransferProgressUpdatingListener and // MultipleFileTransferStateChangeListener downloads.add((DownloadImpl) doDownload( new GetObjectRequest(summary.getBucketName(), summary.getKey()) .withGeneralProgressListener( listener), f, transferListener, null, false)); } if (downloads.isEmpty()) { multipleFileDownload.setState(TransferState.Completed); return multipleFileDownload; } // Notify all state changes waiting for the downloads to all be queued // to wake up and continue. latch.countDown(); return multipleFileDownload; } /** * Uploads all files in the directory given to the bucket named, optionally * recursing for all subdirectories. *

* S3 will overwrite any existing objects that happen to have the same key, * just as when uploading individual files, so use with caution. * * @param bucketName The name of the bucket to upload objects to. * @param virtualDirectoryKeyPrefix The key prefix of the virtual directory * to upload to. Use the null or empty string to upload files to * the root of the bucket. * @param directory The directory to upload. * @param includeSubdirectories Whether to include subdirectories in the * upload. If true, files found in subdirectories will be * included with an appropriate concatenation to the key prefix. */ public MultipleFileUpload uploadDirectory(String bucketName, String virtualDirectoryKeyPrefix, File directory, boolean includeSubdirectories) { return uploadDirectory(bucketName, virtualDirectoryKeyPrefix, directory, includeSubdirectories, null); } /** * Uploads all files in the directory given to the bucket named, optionally * recursing for all subdirectories. *

* S3 will overwrite any existing objects that happen to have the same key, * just as when uploading individual files, so use with caution. * * @param bucketName The name of the bucket to upload objects to. * @param virtualDirectoryKeyPrefix The key prefix of the virtual directory * to upload to. Use the null or empty string to upload files to * the root of the bucket. * @param directory The directory to upload. * @param includeSubdirectories Whether to include subdirectories in the * upload. If true, files found in subdirectories will be * included with an appropriate concatenation to the key prefix. * @param metadataProvider A callback of type * ObjectMetadataProvider which is used to provide * metadata for each file being uploaded. */ public MultipleFileUpload uploadDirectory(String bucketName, String virtualDirectoryKeyPrefix, File directory, boolean includeSubdirectories, ObjectMetadataProvider metadataProvider) { if (directory == null || !directory.exists() || !directory.isDirectory()) { throw new IllegalArgumentException("Must provide a directory to upload"); } final List files = new LinkedList(); listFiles(directory, files, includeSubdirectories); return uploadFileList(bucketName, virtualDirectoryKeyPrefix, directory, files, metadataProvider); } /** * Uploads all specified files to the bucket named, constructing relative * keys depending on the commonParentDirectory given. *

* S3 will overwrite any existing objects that happen to have the same key, * just as when uploading individual files, so use with caution. * * @param bucketName The name of the bucket to upload objects to. * @param virtualDirectoryKeyPrefix The key prefix of the virtual directory * to upload to. Use the null or empty string to upload files to * the root of the bucket. * @param directory The common parent directory of files to upload. The keys * of the files in the list of files are constructed relative to * this directory and the virtualDirectoryKeyPrefix. * @param files A list of files to upload. The keys of the files are * calculated relative to the common parent directory and the * virtualDirectoryKeyPrefix. */ public MultipleFileUpload uploadFileList(String bucketName, String virtualDirectoryKeyPrefix, File directory, List files) { return uploadFileList(bucketName, virtualDirectoryKeyPrefix, directory, files, null); } /** * Uploads all specified files to the bucket named, constructing relative * keys depending on the commonParentDirectory given. *

* S3 will overwrite any existing objects that happen to have the same key, * just as when uploading individual files, so use with caution. * * @param bucketName The name of the bucket to upload objects to. * @param virtualDirectoryKeyPrefix The key prefix of the virtual directory * to upload to. Use the null or empty string to upload files to * the root of the bucket. * @param directory The common parent directory of files to upload. The keys * of the files in the list of files are constructed relative to * this directory and the virtualDirectoryKeyPrefix. * @param files A list of files to upload. The keys of the files are * calculated relative to the common parent directory and the * virtualDirectoryKeyPrefix. * @param metadataProvider A callback of type * ObjectMetadataProvider which is used to provide * metadata for each file being uploaded. */ public MultipleFileUpload uploadFileList(String bucketName, String virtualDirectoryKeyPrefix, File directory, List files, ObjectMetadataProvider metadataProvider) { if (directory == null || !directory.exists() || !directory.isDirectory()) { throw new IllegalArgumentException( "Must provide a common base directory for uploaded files"); } if (virtualDirectoryKeyPrefix == null || virtualDirectoryKeyPrefix.length() == 0) { virtualDirectoryKeyPrefix = ""; } else if (!virtualDirectoryKeyPrefix.endsWith("/")) { virtualDirectoryKeyPrefix = virtualDirectoryKeyPrefix + "/"; } /* This is the hook for adding additional progress listeners */ final ProgressListenerChain additionalListeners = new ProgressListenerChain(); final TransferProgress progress = new TransferProgress(); /* * Bind additional progress listeners to this * MultipleFileTransferProgressUpdatingListener to receive * ByteTransferred events from each single-file upload implementation. */ final ProgressListener listener = new MultipleFileTransferProgressUpdatingListener( progress, additionalListeners); final List uploads = new LinkedList(); final MultipleFileUploadImpl multipleFileUpload = new MultipleFileUploadImpl("Uploading etc", progress, additionalListeners, virtualDirectoryKeyPrefix, bucketName, uploads); multipleFileUpload.setMonitor(new MultipleFileTransferMonitor(multipleFileUpload, uploads)); final CountDownLatch latch = new CountDownLatch(1); final MultipleFileTransferStateChangeListener transferListener = new MultipleFileTransferStateChangeListener( latch, multipleFileUpload); if (files == null || files.isEmpty()) { multipleFileUpload.setState(TransferState.Completed); } else { /* * If the absolute path for the common/base directory does NOT end * in a separator (which is the case for anything but root * directories), then we know there's still a separator between the * base directory and the rest of the file's path, so we increment * the starting position by one. */ int startingPosition = directory.getAbsolutePath().length(); if (!(directory.getAbsolutePath().endsWith(File.separator))) { startingPosition++; } long totalSize = 0; for (final File f : files) { // Check, if file, since only files can be uploaded. if (f.isFile()) { totalSize += f.length(); final String key = f.getAbsolutePath().substring(startingPosition) .replaceAll("\\\\", "/"); final ObjectMetadata metadata = new ObjectMetadata(); // Invoke the callback if it's present. // The callback allows the user to customize the metadata // for each file being uploaded. if (metadataProvider != null) { metadataProvider.provideObjectMetadata(f, metadata); } // All the single-file uploads share the same // MultipleFileTransferProgressUpdatingListener and // MultipleFileTransferStateChangeListener uploads.add((UploadImpl) doUpload( (PutObjectRequest) new PutObjectRequest(bucketName, virtualDirectoryKeyPrefix + key, f) .withMetadata(metadata) .withGeneralProgressListener( listener), transferListener, null, null)); } } progress.setTotalBytesToTransfer(totalSize); } // Notify all state changes waiting for the uploads to all be queued // to wake up and continue latch.countDown(); return multipleFileUpload; } /** * Lists files in the directory given and adds them to the result list * passed in, optionally adding subdirectories recursively. */ private void listFiles(File dir, List results, boolean includeSubDirectories) { final File[] found = dir.listFiles(); if (found != null) { for (final File f : found) { if (f.isDirectory()) { if (includeSubDirectories) { listFiles(f, results, includeSubDirectories); } } else { results.add(f); } } } } /** *

* Aborts any multipart uploads that were initiated before the specified * date. *

*

* This method is useful for cleaning up any interrupted multipart uploads. * TransferManager attempts to abort any failed uploads, but in * some cases this may not be possible, such as if network connectivity is * completely lost. *

* * @param bucketName The name of the bucket containing the multipart uploads * to abort. * @param date The date indicating which multipart uploads should be * aborted. */ public void abortMultipartUploads(String bucketName, Date date) throws AmazonServiceException, AmazonClientException { MultipartUploadListing uploadListing = s3.listMultipartUploads(appendSingleObjectUserAgent( new ListMultipartUploadsRequest(bucketName))); do { for (final MultipartUpload upload : uploadListing.getMultipartUploads()) { if (upload.getInitiated().compareTo(date) < 0) { s3.abortMultipartUpload(appendSingleObjectUserAgent(new AbortMultipartUploadRequest( bucketName, upload.getKey(), upload.getUploadId()))); } } final ListMultipartUploadsRequest request = new ListMultipartUploadsRequest(bucketName) .withUploadIdMarker(uploadListing.getNextUploadIdMarker()) .withKeyMarker(uploadListing.getNextKeyMarker()); uploadListing = s3.listMultipartUploads(appendSingleObjectUserAgent(request)); } while (uploadListing.isTruncated()); } /** * Forcefully shuts down this TransferManager instance - currently executing * transfers will not be allowed to finish. It also by default shuts down * the underlying Amazon S3 client. WARNING: Do not call on the main thread! * * @see #shutdownNow(boolean) */ public void shutdownNow() { shutdownNow(true); } /** * Forcefully shuts down this TransferManager instance - currently executing * transfers will not be allowed to finish. Callers should use this method * when they either: *
    *
  • have already verified that their transfers have completed by checking * each transfer's state *
  • need to exit quickly and don't mind stopping transfers before they * complete. *
*

* Callers should also remember that uploaded parts from an interrupted * upload may not always be automatically cleaned up, but callers can use * {@link #abortMultipartUploads(String, Date)} to clean up any upload * parts. WARNING: Do not call on the main thread! * * @param shutDownS3Client Whether to shut down the underlying Amazon S3 * client. */ public void shutdownNow(boolean shutDownS3Client) { threadPool.shutdownNow(); timedThreadPool.shutdownNow(); if (shutDownS3Client) { if (s3 instanceof AmazonS3Client) { ((AmazonS3Client) s3).shutdown(); } } } /** * Shutdown without interrupting the threads involved, so that, for example, * any upload in progress can complete without throwing * {@link AbortedException}. */ private void shutdown() { threadPool.shutdown(); timedThreadPool.shutdown(); } public static X appendSingleObjectUserAgent(X request) { request.getRequestClientOptions().appendUserAgent(USER_AGENT); return request; } public static X appendMultipartUserAgent(X request) { request.getRequestClientOptions().appendUserAgent(USER_AGENT_MULTIPART); return request; } private static final String USER_AGENT = TransferManager.class.getName() + "/" + VersionInfoUtils.getVersion(); private static final String USER_AGENT_MULTIPART = TransferManager.class.getName() + "_multipart/" + VersionInfoUtils.getVersion(); private static final String DEFAULT_DELIMITER = "/"; /** * There is no need for threads from timedThreadPool if there is no more * running threads in current process, so we need a daemon thread factory * for it. */ private static final ThreadFactory daemonThreadFactory = new ThreadFactory() { final AtomicInteger threadCount = new AtomicInteger(0); @Override public Thread newThread(Runnable r) { final int threadNumber = threadCount.incrementAndGet(); final Thread thread = new Thread(r); thread.setDaemon(true); thread.setName("S3TransferManagerTimedThread-" + threadNumber); return thread; } }; /** *

* Schedules a new transfer to copy data from one Amazon S3 location to * another Amazon S3 location. This method is non-blocking and returns * immediately (i.e. before the copy has finished). *

*

* TransferManager doesn't support copying of encrypted objects * whose encryption materials is stored in instruction file. *

*

* Use the returned Copy object to check if the copy is * complete. *

*

* If resources are available, the copy request will begin immediately. * Otherwise, the copy is scheduled and started as soon as resources become * available. *

* * @param sourceBucketName The name of the bucket from where the object is * to be copied. * @param sourceKey The name of the Amazon S3 object. * @param destinationBucketName The name of the bucket to where the Amazon * S3 object has to be copied. * @param destinationKey The name of the object in the destination bucket. * @return A new Copy object to use to check the state of the * copy request being processed. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Copy copy(String sourceBucketName, String sourceKey, String destinationBucketName, String destinationKey) throws AmazonServiceException, AmazonClientException { return copy(new CopyObjectRequest(sourceBucketName, sourceKey, destinationBucketName, destinationKey)); } /** *

* Schedules a new transfer to copy data from one Amazon S3 location to * another Amazon S3 location. This method is non-blocking and returns * immediately (i.e. before the copy has finished). *

*

* TransferManager doesn't support copying of encrypted objects * whose encryption materials is stored i instruction file. *

*

* Use the returned Copy object to check if the copy is * complete. *

*

* If resources are available, the copy request will begin immediately. * Otherwise, the copy is scheduled and started as soon as resources become * available. *

* * @param copyObjectRequest The request containing all the parameters for * the copy. * @return A new Copy object to use to check the state of the * copy request being processed. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Copy copy(final CopyObjectRequest copyObjectRequest) { return copy(copyObjectRequest, null); } /** *

* Schedules a new transfer to copy data from one Amazon S3 location to * another Amazon S3 location. This method is non-blocking and returns * immediately (i.e. before the copy has finished). *

*

* TransferManager doesn't support copying of encrypted objects * whose encryption materials is stored in instruction file. *

*

* Use the returned Copy object to check if the copy is * complete. *

*

* If resources are available, the copy request will begin immediately. * Otherwise, the copy is scheduled and started as soon as resources become * available. *

* * @param copyObjectRequest The request containing all the parameters for * the copy. * @param stateChangeListener The transfer state change listener to monitor * the copy request * @return A new Copy object to use to check the state of the * copy request being processed. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Copy copy(final CopyObjectRequest copyObjectRequest, final TransferStateChangeListener stateChangeListener) throws AmazonServiceException, AmazonClientException { appendSingleObjectUserAgent(copyObjectRequest); assertParameterNotNull(copyObjectRequest.getSourceBucketName(), "The source bucket name must be specified when a copy request is initiated."); assertParameterNotNull(copyObjectRequest.getSourceKey(), "The source object key must be specified when a copy request is initiated."); assertParameterNotNull( copyObjectRequest.getDestinationBucketName(), "The destination bucket name must be specified when a copy request is initiated."); assertParameterNotNull( copyObjectRequest.getDestinationKey(), "The destination object key must be specified when a copy request is initiated."); final String description = "Copying object from " + copyObjectRequest.getSourceBucketName() + "/" + copyObjectRequest.getSourceKey() + " to " + copyObjectRequest.getDestinationBucketName() + "/" + copyObjectRequest.getDestinationKey(); final GetObjectMetadataRequest getObjectMetadataRequest = new GetObjectMetadataRequest( copyObjectRequest.getSourceBucketName(), copyObjectRequest.getSourceKey()) .withSSECustomerKey(copyObjectRequest.getSourceSSECustomerKey()); final ObjectMetadata metadata = s3.getObjectMetadata(getObjectMetadataRequest); final TransferProgress transferProgress = new TransferProgress(); transferProgress.setTotalBytesToTransfer(metadata.getContentLength()); final ProgressListenerChain listenerChain = new ProgressListenerChain( new TransferProgressUpdatingListener(transferProgress)); final CopyImpl copy = new CopyImpl(description, transferProgress, listenerChain, stateChangeListener); final CopyCallable copyCallable = new CopyCallable(this, threadPool, copy, copyObjectRequest, metadata, listenerChain); final CopyMonitor watcher = new CopyMonitor(this, copy, threadPool, copyCallable, copyObjectRequest, listenerChain); watcher.setTimedThreadPool(timedThreadPool); copy.setMonitor(watcher); return copy; } /** * Resumes an upload operation. This upload operation uses the same * configuration {@link TransferManagerConfiguration} as the original * upload. Any data already uploaded will be skipped, and only the remaining * will be uploaded to Amazon S3. * * @param persistableUpload the upload to resume. * @return A new Upload object to use to check the state of the * upload, listen for progress notifications, and otherwise manage * the upload. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Upload resumeUpload(PersistableUpload persistableUpload) { assertParameterNotNull(persistableUpload, "PauseUpload is mandatory to resume a upload."); configuration.setMinimumUploadPartSize(persistableUpload.getPartSize()); configuration.setMultipartUploadThreshold(persistableUpload .getMutlipartUploadThreshold()); return doUpload(new PutObjectRequest(persistableUpload.getBucketName(), persistableUpload.getKey(), new File(persistableUpload.getFile())), null, null, persistableUpload); } /** * Resumes an download operation. This download operation uses the same * configuration as the original download. Any data already fetched will be * skipped, and only the remaining data is retrieved from Amazon S3. * * @param persistableDownload the download to resume. * @return A new Download object to use to check the state of * the download, listen for progress notifications, and otherwise * manage the download. * @throws AmazonClientException If any errors are encountered in the client * while making the request or handling the response. * @throws AmazonServiceException If any errors occurred in Amazon S3 while * processing the request. */ public Download resumeDownload(PersistableDownload persistableDownload) { assertParameterNotNull(persistableDownload, "PausedDownload is mandatory to resume a download."); final GetObjectRequest request = new GetObjectRequest( persistableDownload.getBucketName(), persistableDownload.getKey(), persistableDownload.getVersionId()); if (persistableDownload.getRange() != null && persistableDownload.getRange().length == 2) { final long[] range = persistableDownload.getRange(); request.setRange(range[0], range[1]); } request.setRequesterPays(persistableDownload.isRequesterPays()); request.setResponseHeaders(persistableDownload.getResponseHeaders()); return doDownload(request, new File(persistableDownload.getFile()), null, null, APPEND_MODE); } /** *

* Asserts that the specified parameter value is not null and * if it is, throws an IllegalArgumentException with the * specified error message. *

* * @param parameterValue The parameter value being checked. * @param errorMessage The error message to include in the * IllegalArgumentException if the specified parameter is null. */ private void assertParameterNotNull(Object parameterValue, String errorMessage) { if (parameterValue == null) { throw new IllegalArgumentException(errorMessage); } } /** * Releasing all resources created by TransferManager before it * is being garbage collected. */ @Override protected void finalize() throws Throwable { shutdown(); } }