com.amazonaws.services.s3.transfer.TransferManager Maven / Gradle / Ivy
Show all versions of aws-java-sdk-s3 Show documentation
/*
* Copyright 2010-2024 Amazon.com, Inc. or its affiliates. All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License").
* You may not use this file except in compliance with the License.
* A copy of the License is located at
*
* http://aws.amazon.com/apache2.0
*
* or in the "license" file accompanying this file. This file is distributed
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
* express or implied. See the License for the specific language governing
* permissions and limitations under the License.
*/
package com.amazonaws.services.s3.transfer;
import static com.amazonaws.services.s3.internal.ServiceUtils.APPEND_MODE;
import static com.amazonaws.services.s3.internal.ServiceUtils.OVERWRITE_MODE;
import com.amazonaws.AbortedException;
import com.amazonaws.AmazonClientException;
import com.amazonaws.AmazonServiceException;
import com.amazonaws.AmazonWebServiceRequest;
import com.amazonaws.SdkClientException;
import com.amazonaws.annotation.SdkInternalApi;
import com.amazonaws.auth.AWSCredentials;
import com.amazonaws.auth.AWSCredentialsProvider;
import com.amazonaws.auth.DefaultAWSCredentialsProviderChain;
import com.amazonaws.event.ProgressListener;
import com.amazonaws.event.ProgressListenerChain;
import com.amazonaws.services.s3.AmazonS3;
import com.amazonaws.services.s3.AmazonS3Client;
import com.amazonaws.services.s3.AmazonS3Encryption;
import com.amazonaws.services.s3.AmazonS3EncryptionV2;
import com.amazonaws.services.s3.internal.FileLocks;
import com.amazonaws.services.s3.internal.Mimetypes;
import com.amazonaws.services.s3.internal.RequestCopyUtils;
import com.amazonaws.services.s3.internal.ServiceUtils;
import com.amazonaws.services.s3.model.AbortMultipartUploadRequest;
import com.amazonaws.services.s3.model.CannedAccessControlList;
import com.amazonaws.services.s3.model.AmazonS3Exception;
import com.amazonaws.services.s3.model.CopyObjectRequest;
import com.amazonaws.services.s3.model.GetObjectMetadataRequest;
import com.amazonaws.services.s3.model.GetObjectRequest;
import com.amazonaws.services.s3.model.ListMultipartUploadsRequest;
import com.amazonaws.services.s3.model.ListObjectsRequest;
import com.amazonaws.services.s3.model.MultipartUpload;
import com.amazonaws.services.s3.model.MultipartUploadListing;
import com.amazonaws.services.s3.model.ObjectListing;
import com.amazonaws.services.s3.model.ObjectMetadata;
import com.amazonaws.services.s3.model.ObjectTagging;
import com.amazonaws.services.s3.model.PresignedUrlDownloadConfig;
import com.amazonaws.services.s3.model.PresignedUrlDownloadRequest;
import com.amazonaws.services.s3.model.PutObjectRequest;
import com.amazonaws.services.s3.model.S3Object;
import com.amazonaws.services.s3.model.S3ObjectInputStream;
import com.amazonaws.services.s3.model.S3ObjectSummary;
import com.amazonaws.services.s3.transfer.Transfer.TransferState;
import com.amazonaws.services.s3.transfer.exception.FileLockException;
import com.amazonaws.services.s3.transfer.internal.CopyCallable;
import com.amazonaws.services.s3.transfer.internal.CopyImpl;
import com.amazonaws.services.s3.transfer.internal.CopyMonitor;
import com.amazonaws.services.s3.transfer.internal.DownloadImpl;
import com.amazonaws.services.s3.transfer.internal.DownloadMonitor;
import com.amazonaws.services.s3.transfer.internal.MultipleFileDownloadImpl;
import com.amazonaws.services.s3.transfer.internal.MultipleFileTransferMonitor;
import com.amazonaws.services.s3.transfer.internal.MultipleFileUploadImpl;
import com.amazonaws.services.s3.transfer.internal.PreparedDownloadContext;
import com.amazonaws.services.s3.transfer.internal.PresignUrlDownloadCallable;
import com.amazonaws.services.s3.transfer.internal.PresignedUrlDownloadImpl;
import com.amazonaws.services.s3.transfer.internal.S3ProgressListener;
import com.amazonaws.services.s3.transfer.internal.S3ProgressListenerChain;
import com.amazonaws.services.s3.transfer.internal.TransferManagerUtils;
import com.amazonaws.services.s3.transfer.internal.TransferProgressUpdatingListener;
import com.amazonaws.services.s3.transfer.internal.TransferStateChangeListener;
import com.amazonaws.services.s3.transfer.internal.UploadCallable;
import com.amazonaws.services.s3.transfer.internal.UploadImpl;
import com.amazonaws.services.s3.transfer.internal.UploadMonitor;
import com.amazonaws.util.IOUtils;
import com.amazonaws.util.VersionInfoUtils;
import java.io.File;
import java.io.IOException;
import java.io.InputStream;
import java.net.URL;
import java.nio.file.Path;
import java.util.ArrayList;
import java.util.Date;
import java.util.LinkedList;
import java.util.List;
import java.util.Stack;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Future;
import java.util.concurrent.ScheduledExecutorService;
import java.util.concurrent.ScheduledThreadPoolExecutor;
import java.util.concurrent.ThreadFactory;
import java.util.concurrent.atomic.AtomicInteger;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
/**
* High level utility for managing transfers to Amazon S3.
*
* TransferManager
provides a simple API for uploading content to
* Amazon S3, and makes extensive use of Amazon S3 multipart uploads to achieve
* enhanced throughput, performance and reliability.
*
* When possible, TransferManager
attempts to use multiple threads
* to upload multiple parts of a single upload at once. When dealing with large
* content sizes and high bandwidth, this can have a significant increase on
* throughput.
*
* TransferManager
is responsible for managing resources such as
* connections and threads; share a single instance of
* TransferManager
whenever possible. TransferManager
,
* like all the client classes in the Amazon Web Services SDK for Java, is thread safe. Call
* TransferManager.shutdownNow()
to release the resources once the
* transfer is complete.
*
* Using TransferManager
to upload objects to Amazon S3 is easy:
*
*
* DefaultAWSCredentialsProviderChain credentialProviderChain = new DefaultAWSCredentialsProviderChain();
* TransferManager tx = new TransferManager(
* credentialProviderChain.getCredentials());
* Upload myUpload = tx.upload(myBucket, myFile.getName(), myFile);
*
* // You can poll your transfer's status to check its progress
* if (myUpload.isDone() == false) {
* System.out.println("Transfer: " + myUpload.getDescription());
* System.out.println(" - State: " + myUpload.getState());
* System.out.println(" - Progress: "
* + myUpload.getProgress().getBytesTransferred());
* }
*
* // Transfers also allow you to set a <code>ProgressListener</code> to receive
* // asynchronous notifications about your transfer's progress.
* myUpload.addProgressListener(myProgressListener);
*
* // Or you can block the current thread and wait for your transfer to
* // to complete. If the transfer fails, this method will throw an
* // AmazonClientException or AmazonServiceException detailing the reason.
* myUpload.waitForCompletion();
*
* // After the upload is complete, call shutdownNow to release the resources.
* tx.shutdownNow();
*
*
* Transfers can be paused and resumed at a later time. It can also survive JVM
* crash, provided the information that is required to resume the transfer is
* given as input to the resume operation. For more information on pause and resume.
*
*
* Migrating to the AWS SDK for Java v2
*
* The v2 equivalent of this class is
* S3TransferManager
*
*
* See Migration Guide
* for more information.
*
* @see Upload#pause()
* @see Download#pause()
* @see TransferManager#resumeUpload(PersistableUpload)
* @see TransferManager#resumeDownload(PersistableDownload)
*/
public class TransferManager {
/** The low level client we use to make the actual calls to Amazon S3. */
private final AmazonS3 s3;
/** Configuration for how TransferManager processes requests. */
private TransferManagerConfiguration configuration;
/** The thread pool in which transfers are uploaded or downloaded. */
private final ExecutorService executorService;
/**
* Thread used for periodically checking transfers and updating their state, as well as enforcing
* timeouts.
*/
private final ScheduledExecutorService timedThreadPool = new ScheduledThreadPoolExecutor(1, daemonThreadFactory);
private static final Log log = LogFactory.getLog(TransferManager.class);
private final boolean shutDownThreadPools;
/**
* Flag indicating whether the transfer manager is mutable or not. Legacy managers built via the
* constructors are mutable. TransferManagers built with the fluent builders are immutable.
*/
private final boolean isImmutable;
/**
* Constructs a new TransferManager
and Amazon S3 client using
* the credentials from DefaultAWSCredentialsProviderChain
*
* TransferManager
and client objects may pool connections and
* threads. Reuse TransferManager
and client objects and share
* them throughout applications.
*
* TransferManager and all Amazon Web Services client objects are thread safe.
* @deprecated use {@link TransferManagerBuilder#defaultTransferManager()}
*/
@Deprecated
public TransferManager(){
this(new AmazonS3Client(new DefaultAWSCredentialsProviderChain()));
}
/**
* Constructs a new TransferManager
and Amazon S3 client using
* the specified Amazon Web Services security credentials provider.
*
* TransferManager
and client objects may pool connections and
* threads. Reuse TransferManager
and client objects and share
* them throughout applications.
*
* TransferManager and all Amazon Web Services client objects are thread safe.
*
* @param credentialsProvider
* The Amazon Web Services security credentials provider to use when making
* authenticated requests.
* @deprecated use {@link TransferManagerBuilder#withS3Client(AmazonS3)} for example:
* {@code TransferManagerBuilder.standard().withS3Client(AmazonS3ClientBuilder.standard.withCredentials(credentialsProvider).build()).build(); }
*/
@Deprecated
public TransferManager(AWSCredentialsProvider credentialsProvider) {
this(new AmazonS3Client(credentialsProvider));
}
/**
* Constructs a new TransferManager
and Amazon S3 client using
* the specified Amazon Web Services security credentials.
*
* TransferManager
and client objects
* may pool connections and threads.
* Reuse TransferManager
and client objects
* and share them throughout applications.
*
* TransferManager and all Amazon Web Services client objects are thread safe.
*
* @param credentials
* The Amazon Web Services security credentials to use when making authenticated
* requests.
* @deprecated use {@link TransferManagerBuilder#withS3Client(AmazonS3)} for example:
* {@code TransferManagerBuilder.standard().withS3Client(AmazonS3ClientBuilder.standard.withCredentials(credentials).build()).build(); }
*/
@Deprecated
public TransferManager(AWSCredentials credentials) {
this(new AmazonS3Client(credentials));
}
/**
* Constructs a new TransferManager
,
* specifying the client to use when making
* requests to Amazon S3.
*
* TransferManager
and client objects
* may pool connections and threads.
* Reuse TransferManager
and client objects
* and share them throughout applications.
*
* TransferManager and all Amazon Web Services client objects are thread safe.
*
*
* @param s3
* The client to use when making requests to Amazon S3.
* @deprecated use {@link TransferManagerBuilder#withS3Client(AmazonS3)}
*/
@Deprecated
public TransferManager(AmazonS3 s3) {
this(s3, TransferManagerUtils.createDefaultExecutorService());
}
/**
* Constructs a new TransferManager
specifying the client and
* thread pool to use when making requests to Amazon S3.
*
* TransferManager
and client objects may pool connections and
* threads. Reuse TransferManager
and client objects and share
* them throughout applications.
*
* TransferManager and all Amazon Web Services client objects are thread safe.
*
* By default, the thread pool will shutdown when the transfer manager
* instance is garbage collected.
*
* @param s3
* The client to use when making requests to Amazon S3.
* @param executorService
* The ExecutorService to use for the TransferManager. It is not recommended to
* use a single threaded executor or a thread pool with a bounded work queue as
* control tasks may submit subtasks that can't complete until all sub tasks
* complete. Using an incorrectly configured thread pool may cause a deadlock (I.E.
* the work queue is filled with control tasks that can't finish until subtasks
* complete but subtasks can't execute because the queue is filled).
*
* @see TransferManager#TransferManager(AmazonS3 s3, ExecutorService
* executorService, boolean shutDownThreadPools)
* @deprecated use {@link TransferManagerBuilder#withS3Client(AmazonS3)} and
* {@link TransferManagerBuilder#withExecutorFactory(com.amazonaws.client.builder.ExecutorFactory)}
*/
@Deprecated
public TransferManager(AmazonS3 s3, ExecutorService executorService) {
this(s3, executorService, true);
}
/**
* Constructs a new TransferManager
specifying the client and
* thread pool to use when making requests to Amazon S3.
*
* TransferManager
and client objects may pool connections and
* threads. Reuse TransferManager
and client objects and share
* them throughout applications.
*
* TransferManager and all Amazon Web Services client objects are thread safe.
*
* @param s3
* The client to use when making requests to Amazon S3.
* @param executorService
* The ExecutorService to use for the TransferManager. It is not recommended to
* use a single threaded executor or a thread pool with a bounded work queue as
* control tasks may submit subtasks that can't complete until all sub tasks
* complete. Using an incorrectly configured thread pool may cause a deadlock (I.E.
* the work queue is filled with control tasks that can't finish until subtasks
* complete but subtasks can't execute because the queue is filled).
* @param shutDownThreadPools
* If set to true, the thread pool will be shutdown when transfer
* manager instance is garbage collected.
* @deprecated use {@link TransferManagerBuilder#withS3Client(AmazonS3)} and
* {@link TransferManagerBuilder#withExecutorFactory(com.amazonaws.client.builder.ExecutorFactory)} and
* {@link TransferManagerBuilder#withShutDownThreadPools(Boolean)}
*/
@Deprecated
public TransferManager(AmazonS3 s3, ExecutorService executorService, boolean shutDownThreadPools) {
this.s3 = s3;
this.executorService = executorService;
this.configuration = new TransferManagerConfiguration();
this.shutDownThreadPools = shutDownThreadPools;
this.isImmutable = false;
}
@SdkInternalApi
TransferManager(TransferManagerParams params) {
this.s3 = params.getS3Client();
this.executorService = params.getExecutorService();
this.configuration = params.getConfiguration();
this.shutDownThreadPools = params.getShutDownThreadPools();
this.isImmutable = true;
}
/**
* Constructor for use by classes that need to extend the TransferManager.
*
* @param builder - transfer manager builder with the required configuration
*/
protected TransferManager(TransferManagerBuilder builder) {
this(builder.getParams());
}
/**
* Sets the configuration which specifies how
* this TransferManager
processes requests.
*
* @param configuration
* The new configuration specifying how
* this TransferManager
* processes requests.
* @deprecated use appropriate method on the {@link TransferManagerBuilder} for example:
* {@code TransferManagerBuilder.standard().withMinimumUploadPartSize(100L).build(); }
*/
@Deprecated
public void setConfiguration(TransferManagerConfiguration configuration) {
checkMutability();
this.configuration = configuration;
}
/**
* Returns the configuration which specifies how
* this TransferManager
processes requests.
*
* @return The configuration settings for this TransferManager
.
*/
public TransferManagerConfiguration getConfiguration() {
return configuration;
}
/**
* Returns the underlying Amazon S3 client used to make requests to
* Amazon S3.
*
* @return The underlying Amazon S3 client used to make requests to
* Amazon S3.
*/
public AmazonS3 getAmazonS3Client() {
return s3;
}
/**
*
* Schedules a new transfer to upload data to Amazon S3. This method is
* non-blocking and returns immediately (i.e. before the upload has
* finished).
*
*
* When uploading options from a stream, callers must supply the size of
* options in the stream through the content length field in the
* ObjectMetadata
parameter.
* If no content length is specified for the input
* stream, then TransferManager will attempt to buffer all the stream
* contents in memory and upload the options as a traditional, single part
* upload. Because the entire stream contents must be buffered in memory,
* this can be very expensive, and should be avoided whenever possible.
*
*
* Use the returned Upload
object to query the progress of the
* transfer, add listeners for progress events, and wait for the upload to
* complete.
*
*
* If resources are available, the upload will begin immediately.
* Otherwise, the upload is scheduled and started as soon as
* resources become available.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload the new object to.
* @param key
* The key in the specified bucket by which to store the new
* object.
* @param input
* The input stream containing the options to upload to Amazon S3.
* @param objectMetadata
* Additional information about the object being uploaded,
* including the size of the options, content type, additional
* custom user metadata, etc.
*
* @return A new Upload
object to use to check
* the state of the upload, listen for progress notifications,
* and otherwise manage the upload.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Upload upload(final String bucketName, final String key, final InputStream input, ObjectMetadata objectMetadata)
throws AmazonServiceException, AmazonClientException {
return upload(new PutObjectRequest(bucketName, key, input, objectMetadata));
}
/**
* Schedules a new transfer to upload data to Amazon S3. This method is
* non-blocking and returns immediately (i.e. before the upload has
* finished).
*
* The returned Upload object allows you to query the progress of the
* transfer, add listeners for progress events, and wait for the upload to
* complete.
*
* If resources are available, the upload will begin immediately, otherwise
* it will be scheduled and started as soon as resources become available.
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload the new object to.
* @param key
* The key in the specified bucket by which to store the new
* object.
* @param file
* The file to upload.
*
* @return A new Upload object which can be used to check state of the
* upload, listen for progress notifications, and otherwise manage
* the upload.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Upload upload(final String bucketName, final String key, final File file)
throws AmazonServiceException, AmazonClientException {
return upload(new PutObjectRequest(bucketName, key, file));
}
/**
*
* Schedules a new transfer to upload data to Amazon S3. This method is
* non-blocking and returns immediately (i.e. before the upload has
* finished).
*
*
* Use the returned Upload
object to query the progress of the
* transfer, add listeners for progress events, and wait for the upload to
* complete.
*
*
* If resources are available, the upload will begin immediately.
* Otherwise, the upload is scheduled and started as soon as
* resources become available.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param putObjectRequest
* The request containing all the parameters for the upload.
*
* @return A new Upload
object to use to check
* the state of the upload, listen for progress notifications,
* and otherwise manage the upload.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Upload upload(final PutObjectRequest putObjectRequest)
throws AmazonServiceException, AmazonClientException {
return doUpload(putObjectRequest, null, null, null);
}
/**
*
* Schedules a new transfer to upload data to Amazon S3. This method is
* non-blocking and returns immediately (i.e. before the upload has
* finished).
*
*
* Use the returned Upload
object to query the progress of the
* transfer, add listeners for progress events, and wait for the upload to
* complete.
*
*
* If resources are available, the upload will begin immediately. Otherwise,
* the upload is scheduled and started as soon as resources become
* available.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param putObjectRequest
* The request containing all the parameters for the upload.
* @param progressListener
* An optional callback listener to receive the progress of the
* upload.
*
* @return A new Upload
object to use to check the state of the
* upload, listen for progress notifications, and otherwise manage
* the upload.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Upload upload(final PutObjectRequest putObjectRequest,
final S3ProgressListener progressListener)
throws AmazonServiceException, AmazonClientException {
return doUpload(putObjectRequest, null, progressListener, null);
}
/**
*
* Schedules a new transfer to upload data to Amazon S3. This method is
* non-blocking and returns immediately (i.e. before the upload has
* finished).
*
*
* Use the returned Upload
object to query the progress of the
* transfer, add listeners for progress events, and wait for the upload to
* complete.
*
*
* If resources are available, the upload will begin immediately. Otherwise,
* the upload is scheduled and started as soon as resources become
* available.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param putObjectRequest
* The request containing all the parameters for the upload.
* @param stateListener
* The transfer state change listener to monitor the upload.
* @param progressListener
* An optional callback listener to receive the progress of the
* upload.
*
* @return A new Upload
object to use to check the state of the
* upload, listen for progress notifications, and otherwise manage
* the upload.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
private Upload doUpload(final PutObjectRequest putObjectRequest,
final TransferStateChangeListener stateListener,
final S3ProgressListener progressListener,
final PersistableUpload persistableUpload) throws AmazonServiceException,
AmazonClientException {
assertNotObjectLambdaArn(putObjectRequest.getBucketName(), "upload");
appendSingleObjectUserAgent(putObjectRequest);
String multipartUploadId = persistableUpload != null ? persistableUpload
.getMultipartUploadId() : null;
if (putObjectRequest.getMetadata() == null)
putObjectRequest.setMetadata(new ObjectMetadata());
ObjectMetadata metadata = putObjectRequest.getMetadata();
File file = TransferManagerUtils.getRequestFile(putObjectRequest);
if ( file != null ) {
// Always set the content length, even if it's already set
metadata.setContentLength(file.length());
// Only set the content type if it hasn't already been set
if ( metadata.getContentType() == null ) {
metadata.setContentType(Mimetypes.getInstance().getMimetype(file));
}
} else {
if (multipartUploadId != null) {
throw new IllegalArgumentException(
"Unable to resume the upload. No file specified.");
}
}
String description = "Uploading to " + putObjectRequest.getBucketName()
+ "/" + putObjectRequest.getKey();
TransferProgress transferProgress = new TransferProgress();
transferProgress.setTotalBytesToTransfer(TransferManagerUtils
.getContentLength(putObjectRequest));
S3ProgressListenerChain listenerChain = new S3ProgressListenerChain(
new TransferProgressUpdatingListener(transferProgress),
putObjectRequest.getGeneralProgressListener(), progressListener);
putObjectRequest.setGeneralProgressListener(listenerChain);
UploadImpl upload = new UploadImpl(description, transferProgress,
listenerChain, stateListener);
/**
* Since we use the same thread pool for uploading individual parts and
* complete multi part upload, there is a possibility that the tasks for
* complete multi-part upload will be added to end of queue in case of
* multiple parallel uploads submitted. This may result in a delay for
* processing the complete multi part upload request.
*/
UploadCallable uploadCallable = new UploadCallable(this, executorService,
upload, putObjectRequest, listenerChain, multipartUploadId,
transferProgress);
UploadMonitor watcher = UploadMonitor.create(this, upload, executorService,
uploadCallable, putObjectRequest, listenerChain);
upload.setMonitor(watcher);
return upload;
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucket
* The name of the bucket containing the object to download.
* @param key
* The key under which the object to download is stored.
* @param file
* The file to download the object's data to.
*
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(String bucket, String key, File file) {
return download(bucket, key, file, 0);
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucket
* The name of the bucket containing the object to download.
* @param key
* The key under which the object to download is stored.
* @param file
* The file to download the object's data to.
* @param timeoutMillis
* Timeout, in milliseconds, for waiting for this download to
* complete. Note that the timeout time will be approximate
* and is not strictly guaranteed. As a result this timeout
* should not be relied on in cases where exact precision is
* required.
*
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(String bucket, String key,
File file, long timeoutMillis) {
return download(new GetObjectRequest(bucket, key), file,
timeoutMillis);
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param getObjectRequest
* The request containing all the parameters for the download.
* @param file
* The file to download the object data to.
*
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(final GetObjectRequest getObjectRequest, final File file) {
return download(getObjectRequest, file, 0);
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param getObjectRequest
* The request containing all the parameters for the download.
* @param file
* The file to download the object data to.
* @param timeoutMillis
* Timeout, in milliseconds, for waiting for this download to
* complete. Note that the timeout time will be approximate
* and is not strictly guaranteed. As a result this timeout
* should not be relied on in cases where exact precision is
* required.
*
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(final GetObjectRequest getObjectRequest,
final File file, long timeoutMillis) {
return doDownload(getObjectRequest, file, null, null, OVERWRITE_MODE,
timeoutMillis, null);
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param getObjectRequest
* The request containing all the parameters for the download.
* @param file
* The file to download the object data to.
* @param progressListener
* An optional callback listener to get the progress of the
* download.
*
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(final GetObjectRequest getObjectRequest,
final File file, final S3ProgressListener progressListener) {
return doDownload(getObjectRequest, file, null, progressListener,
OVERWRITE_MODE, 0, null);
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param getObjectRequest
* The request containing all the parameters for the download.
* @param file
* The file to download the object data to.
* @param progressListener
* An optional callback listener to get the progress of the
* download.
* @param timeoutMillis
* Timeout, in milliseconds, for waiting for this download to
* complete. Note that the timeout time will be approximate
* and is not strictly guaranteed. As a result this timeout
* should not be relied on in cases where exact precision is
* required.
*
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(final GetObjectRequest getObjectRequest,
final File file, final S3ProgressListener progressListener,
final long timeoutMillis) {
return doDownload(getObjectRequest, file, null, progressListener,
OVERWRITE_MODE, timeoutMillis, null);
}
/**
* Schedules a new transfer to download data from Amazon S3 and save it to
* the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned Download object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param getObjectRequest
* The request containing all the parameters for the download.
* @param file
* The file to download the object data to.
* @param progressListener
* An optional callback listener to get the progress of the
* download.
* @param timeoutMillis
* Timeout, in milliseconds, for waiting for this download to
* complete. Note that the timeout time will be approximate
* and is not strictly guaranteed. As a result this timeout
* should not be relied on in cases where exact precision is
* required.
* @param resumeOnRetry
* If set to true, upon an immediate retry of a failed object
* download, the TransferManager
will resume the
* download from the current end of the file on disk.
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download download(final GetObjectRequest getObjectRequest,
final File file, final S3ProgressListener progressListener,
final long timeoutMillis, final boolean resumeOnRetry) {
return doDownload(getObjectRequest, file, null, progressListener,
OVERWRITE_MODE, timeoutMillis, null, 0L, resumeOnRetry, 0L);
}
private Download doDownload(final GetObjectRequest getObjectRequest,
final File file, final TransferStateChangeListener stateListener,
final S3ProgressListener s3progressListener,
final boolean resumeExistingDownload,
final long timeoutMillis,
final PersistableDownload persistableDownload) {
assertNotObjectLambdaArn(getObjectRequest.getBucketName(), "download");
long lastModifiedTimeRecordedDuringPause = 0L;
Integer lastFullyDownloadedPartNumber = null;
Long lastFullyDownloadedFilePosition = null;
if (persistableDownload != null) {
lastModifiedTimeRecordedDuringPause = persistableDownload.getlastModifiedTime();
lastFullyDownloadedPartNumber = persistableDownload.getLastFullyDownloadedPartNumber();
lastFullyDownloadedFilePosition = persistableDownload.getLastFullyDownloadedFilePosition();
}
return doDownload(getObjectRequest, file, stateListener, s3progressListener,
resumeExistingDownload, timeoutMillis, lastFullyDownloadedPartNumber,
lastModifiedTimeRecordedDuringPause, false, lastFullyDownloadedFilePosition);
}
/**
* Same as public interface, but adds a state listener so that callers can
* be notified of state changes to the download.
*
* @see TransferManager#download(GetObjectRequest, File)
*/
private Download doDownload(final GetObjectRequest getObjectRequest,
final File file, final TransferStateChangeListener stateListener,
final S3ProgressListener s3progressListener,
final boolean resumeExistingDownload,
final long timeoutMillis,
final Integer lastFullyDownloadedPart,
final long lastModifiedTimeRecordedDuringPause,
final boolean resumeOnRetry,
final Long lastFullyDownloadedPartPosition) {
assertNotObjectLambdaArn(getObjectRequest.getBucketName(), "download");
PreparedDownloadContext prepared = prepareDownload(getObjectRequest,
file,
stateListener,
s3progressListener,
resumeExistingDownload,
timeoutMillis,
lastFullyDownloadedPart,
lastModifiedTimeRecordedDuringPause,
resumeOnRetry,
lastFullyDownloadedPartPosition);
return submitDownload(prepared);
}
private PreparedDownloadContext prepareDownload(final GetObjectRequest getObjectRequest,
final File file, final TransferStateChangeListener stateListener,
final S3ProgressListener s3progressListener,
final boolean resumeExistingDownload,
final long timeoutMillis,
final Integer lastFullyDownloadedPart,
final long lastModifiedTimeRecordedDuringPause,
final boolean resumeOnRetry,
final Long lastFullyDownloadedPartPosition) {
assertParameterNotNull(getObjectRequest,
"A valid GetObjectRequest must be provided to initiate download");
assertParameterNotNull(file,
"A valid file must be provided to download into");
appendSingleObjectUserAgent(getObjectRequest);
String description = "Downloading from " + getObjectRequest.getBucketName() + "/" + getObjectRequest.getKey();
TransferProgress transferProgress = new TransferProgress();
// S3 progress listener to capture the persistable transfer when available
S3ProgressListenerChain listenerChain = new S3ProgressListenerChain(
// The listener for updating transfer progress
new TransferProgressUpdatingListener(transferProgress),
getObjectRequest.getGeneralProgressListener(),
s3progressListener); // Listeners included in the original request
// The listener chain used by the low-level GetObject request.
// This listener chain ignores any COMPLETE event, so that we could
// delay firing the signal until the high-level download fully finishes.
getObjectRequest
.setGeneralProgressListener(new ProgressListenerChain(new TransferCompletionFilter(), listenerChain));
GetObjectMetadataRequest getObjectMetadataRequest = RequestCopyUtils.createGetObjectMetadataRequestFrom(getObjectRequest);
final ObjectMetadata objectMetadata = s3.getObjectMetadata(getObjectMetadataRequest);
// Used to check if the object is modified between pause and resume
long lastModifiedTime = objectMetadata.getLastModified().getTime();
long startingByte = 0;
long lastByte;
long[] range = getObjectRequest.getRange();
if (range != null && range.length == 2) {
startingByte = range[0];
lastByte = range[1];
} else {
lastByte = objectMetadata.getContentLength() - 1;
}
final long origStartingByte = startingByte;
final boolean isDownloadParallel = !configuration.isDisableParallelDownloads()
&& TransferManagerUtils.isDownloadParallelizable(s3, getObjectRequest, ServiceUtils.getPartCount(getObjectRequest, s3));
// We still pass the unfiltered listener chain into DownloadImpl
final DownloadImpl download = new DownloadImpl(description, transferProgress, listenerChain, null,
stateListener, getObjectRequest, file, objectMetadata, isDownloadParallel);
long totalBytesToDownload = lastByte - startingByte + 1;
transferProgress.setTotalBytesToTransfer(totalBytesToDownload);
// Range information is needed for auto retry of downloads so a retry
// request can start at the last downloaded location in the range.
//
// For obvious reasons, setting a Range header only makes sense if the
// object actually has content because it's inclusive, otherwise S3
// responds with 4xx
//
// In addition, we only set the range if the download was *NOT*
// determined to be parallelizable above. One of the conditions for
// parallel downloads is that getRange() returns null so preserve that.
if (totalBytesToDownload > 0 && !isDownloadParallel) {
getObjectRequest.withRange(startingByte, lastByte);
}
long fileLength = -1;
if (resumeExistingDownload) {
if (isS3ObjectModifiedSincePause(lastModifiedTime, lastModifiedTimeRecordedDuringPause)) {
throw new AmazonClientException("The requested object in bucket " + getObjectRequest.getBucketName()
+ " with key " + getObjectRequest.getKey() + " is modified on Amazon S3 since the last pause.");
}
// There's still a chance the object is modified while the request
// is in flight. Set this header so S3 fails the request if this happens.
getObjectRequest.setUnmodifiedSinceConstraint(new Date(lastModifiedTime));
if (!isDownloadParallel) {
if (!FileLocks.lock(file)) {
throw new FileLockException("Fail to lock " + file + " for resume download");
}
try {
if (file.exists()) {
fileLength = file.length();
startingByte = startingByte + fileLength;
getObjectRequest.setRange(startingByte, lastByte);
transferProgress.updateProgress(Math.min(fileLength, totalBytesToDownload));
totalBytesToDownload = lastByte - startingByte + 1;
if (log.isDebugEnabled()) {
log.debug("Resume download: totalBytesToDownload=" + totalBytesToDownload
+ ", origStartingByte=" + origStartingByte + ", startingByte=" + startingByte
+ ", lastByte=" + lastByte + ", numberOfBytesRead=" + fileLength + ", file: "
+ file);
}
}
} finally {
FileLocks.unlock(file);
}
}
}
if (totalBytesToDownload < 0) {
throw new IllegalArgumentException(
"Unable to determine the range for download operation.");
}
final CountDownLatch latch = new CountDownLatch(1);
DownloadCallable downloadCallable = new DownloadCallable(s3, latch,
getObjectRequest, resumeExistingDownload,
download, file, origStartingByte, fileLength, timeoutMillis, timedThreadPool,
executorService, lastFullyDownloadedPart, isDownloadParallel, resumeOnRetry)
.withLastFullyMergedPartPosition(lastFullyDownloadedPartPosition);
return new PreparedDownloadContext(download, downloadCallable, latch);
}
private DownloadImpl submitDownload(PreparedDownloadContext preparedDownloadContext) {
Future future = executorService.submit(preparedDownloadContext.getCallable());
DownloadImpl transfer = preparedDownloadContext.getTransfer();
transfer.setMonitor(new DownloadMonitor(transfer, future));
preparedDownloadContext.getLatch().countDown();
return transfer;
}
private boolean isS3ObjectModifiedSincePause(final long lastModifiedTimeRecordedDuringResume,
long lastModifiedTimeRecordedDuringPause) {
return lastModifiedTimeRecordedDuringResume != lastModifiedTimeRecordedDuringPause;
}
/**
* Schedules a new transfer to download data from Amazon S3 using presigned url
* and save it to the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned {@link PresignedUrlDownload} object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* Note: The result of the operation doesn't support pause and resume functionality.
*
* @param request The request containing all the parameters for the download.
* @param destFile The file to download the object data to.
*
* @return A new {@link PresignedUrlDownload} object to check the state of the download,
* listen for progress notifications, and otherwise manage the download.
*/
public PresignedUrlDownload download(final PresignedUrlDownloadRequest request, final File destFile) {
return download(request, destFile, new PresignedUrlDownloadConfig());
}
/**
* Schedules a new transfer to download data from Amazon S3 using presigned url
* and save it to the specified file. This method is non-blocking and returns immediately
* (i.e. before the data has been fully downloaded).
*
* Use the returned {@link PresignedUrlDownload} object to query the progress of the transfer,
* add listeners for progress events, and wait for the download to complete.
*
*
* Note: The result of the operation doesn't support pause and resume functionality.
*
*
* @param request The request containing all the parameters for the download.
* @param destFile The file to download the object data to.
* @param downloadContext Additional configuration to control the download behavior
*
* @return A new {@link PresignedUrlDownload} object to check the state of the download,
* listen for progress notifications, and otherwise manage the download.
*/
public PresignedUrlDownload download(final PresignedUrlDownloadRequest request,
final File destFile,
final PresignedUrlDownloadConfig downloadContext) {
assertParameterNotNull(request,
"A valid PresignedUrlDownloadRequest must be provided to initiate download");
assertParameterNotNull(destFile,
"A valid file must be provided to download into");
assertParameterNotNull(downloadContext,
"A valid PresignedUrlDownloadContext must be provided");
assertNotObjectLambdaUrl(request.getPresignedUrl(), "download");
appendSingleObjectUserAgent(request);
String description = "Downloading from the given presigned url: " + request.getPresignedUrl();
TransferProgress transferProgress = new TransferProgress();
S3ProgressListenerChain listenerChain = new S3ProgressListenerChain(new TransferProgressUpdatingListener(transferProgress),
request.getGeneralProgressListener(),
downloadContext.getS3progressListener());
request.setGeneralProgressListener(new ProgressListenerChain(new TransferCompletionFilter(), listenerChain));
Long startByte = 0L;
Long endByte = null;
long[] range = request.getRange();
if (range != null && range.length == 2) {
startByte = range[0];
endByte = range[1];
} else {
// Get content length by making a range GET call
final ObjectMetadata objectMetadata = getObjectMetadataUsingRange(request);
if (objectMetadata != null) {
Long contentLength = TransferManagerUtils.getContentLengthFromContentRange(objectMetadata);
endByte = contentLength != null ? contentLength - 1 : null;
}
}
final long perRequestDownloadSize = downloadContext.getDownloadSizePerRequest();
final boolean isDownloadParallel = isDownloadParallel(request, startByte, endByte, perRequestDownloadSize);
final PresignedUrlDownloadImpl download =
new PresignedUrlDownloadImpl(description, transferProgress, listenerChain, request);
if (startByte != null && endByte != null) {
transferProgress.setTotalBytesToTransfer(endByte - startByte + 1);
}
final CountDownLatch latch = new CountDownLatch(1);
Future future = executorService.submit(
new PresignUrlDownloadCallable(executorService, destFile, latch, download, isDownloadParallel, timedThreadPool,
downloadContext.getTimeoutMillis(), s3, request, perRequestDownloadSize,
startByte, endByte, downloadContext.isResumeOnRetry()));
download.setMonitor(new DownloadMonitor(download, future));
latch.countDown();
return download;
}
/**
* Returns the object metadata using the {@link PresignedUrlDownloadRequest}.
* Returns null if object size is zero.
*/
private ObjectMetadata getObjectMetadataUsingRange(final PresignedUrlDownloadRequest request) {
PresignedUrlDownloadRequest copy = request.clone();
S3Object s3Object = null;
try {
s3Object = s3.download(copy.withRange(0, 0))
.getS3Object();
return s3Object.getObjectMetadata();
} catch (AmazonS3Exception exception) {
// This handles error case when trying a range GET on object with zero size
if (exception.getStatusCode() == 416 && "InvalidRange".equals(exception.getErrorCode())) {
return null;
}
throw exception;
} finally {
if (s3Object != null) {
S3ObjectInputStream objectContent = s3Object.getObjectContent();
IOUtils.drainInputStream(objectContent);
IOUtils.closeQuietly(objectContent, log);
}
}
}
/**
* Returns a boolean value indicating if object can be downloaded in parallel
* when using presigned url.
*/
private boolean isDownloadParallel(PresignedUrlDownloadRequest request, Long startByte, Long endByte,
long partialObjectMaxSize) {
return !configuration.isDisableParallelDownloads() && !(s3 instanceof AmazonS3Encryption)
&& !(s3 instanceof AmazonS3EncryptionV2)
// Can't rely on set range as endbyte can be set to random number longer than actual size. This results in
// making large number of partial requests even after the entire file is read from S3
&& request.getRange() == null
// This is when SDK can compute the endByte through the object metadata
&& (startByte != null && endByte != null && endByte - startByte + 1 > partialObjectMaxSize);
}
public MultipleFileDownload downloadDirectory(String bucketName, String keyPrefix, File destinationDirectory) {
return downloadDirectory(bucketName, keyPrefix, destinationDirectory, false);
}
public MultipleFileDownload downloadDirectory(String bucketName, String keyPrefix, File destinationDirectory, KeyFilter filter) {
return downloadDirectory(bucketName, keyPrefix, destinationDirectory, false, filter);
}
public MultipleFileDownload downloadDirectory(String bucketName, String keyPrefix, File destinationDirectory,
boolean resumeOnRetry) {
return downloadDirectory(bucketName, keyPrefix, destinationDirectory, resumeOnRetry, null);
}
/**
* Downloads all objects in the virtual directory designated by the
* keyPrefix given to the destination directory given. All virtual
* subdirectories will be downloaded recursively.
*
* If you are downloading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The bucket containing the virtual directory
* @param keyPrefix
* The key prefix for the virtual directory, or null for the
* entire bucket. All subdirectories will be downloaded
* recursively.
* @param destinationDirectory
* The directory to place downloaded files. Subdirectories will
* be created as necessary.
* @param resumeOnRetry
* If set to true, upon an immediate retry of a failed object
* download, the TransferManager
will resume the
* download from the current end of the file on disk.
* @param filter
* If set, applies the filter to determine which keys to include
* in the download request. (default is include all).
*/
public MultipleFileDownload downloadDirectory(String bucketName, String keyPrefix, File destinationDirectory,
boolean resumeOnRetry, KeyFilter filter) {
assertNotObjectLambdaArn(bucketName, "downloadDirectory");
if ( keyPrefix == null )
keyPrefix = "";
if ( filter == null ) {
filter = KeyFilter.INCLUDE_ALL;
}
List objectSummaries = new LinkedList();
Stack commonPrefixes = new Stack();
commonPrefixes.add(keyPrefix);
long totalSize = 0;
// Recurse all virtual subdirectories to get a list of object summaries.
// This is a depth-first search.
do {
String prefix = commonPrefixes.pop();
ObjectListing listObjectsResponse = null;
do {
if ( listObjectsResponse == null ) {
ListObjectsRequest listObjectsRequest = new ListObjectsRequest().withBucketName(bucketName)
.withDelimiter(DEFAULT_DELIMITER).withPrefix(prefix);
listObjectsResponse = s3.listObjects(listObjectsRequest);
} else {
listObjectsResponse = s3.listNextBatchOfObjects(listObjectsResponse);
}
for ( S3ObjectSummary s : listObjectsResponse.getObjectSummaries() ) {
// Skip any files that are also virtual directories, since
// we can't save both a directory and a file of the same
// name.
if ( !filter.shouldInclude(s) ) {
log.debug("Skipping " + s.getKey() + " as it does not match filter.");
continue;
}
if ( leavesRoot(destinationDirectory, s.getKey()) ) {
throw new RuntimeException("Cannot download key " + s.getKey() +
", its relative path resolves outside the parent directory.");
}
if ( !s.getKey().equals(prefix)
&& !listObjectsResponse.getCommonPrefixes().contains(s.getKey() + DEFAULT_DELIMITER) ) {
objectSummaries.add(s);
totalSize += s.getSize();
} else {
log.debug("Skipping download for object " + s.getKey()
+ " since it is also a virtual directory");
}
}
commonPrefixes.addAll(listObjectsResponse.getCommonPrefixes());
} while ( listObjectsResponse.isTruncated() );
} while ( !commonPrefixes.isEmpty() );
/* This is the hook for adding additional progress listeners */
ProgressListenerChain additionalListeners = new ProgressListenerChain();
TransferProgress transferProgress = new TransferProgress();
transferProgress.setTotalBytesToTransfer(totalSize);
/*
* Bind additional progress listeners to this
* MultipleFileTransferProgressUpdatingListener to receive
* ByteTransferred events from each single-file download implementation.
*/
ProgressListener listener = new MultipleFileTransferProgressUpdatingListener(
transferProgress, additionalListeners);
List downloads = new ArrayList();
List preparedDownloadContexts = new ArrayList();
String description = "Downloading from " + bucketName + "/" + keyPrefix;
final MultipleFileDownloadImpl multipleFileDownload = new MultipleFileDownloadImpl(description, transferProgress,
additionalListeners, keyPrefix, bucketName, downloads);
multipleFileDownload.setMonitor(new MultipleFileTransferMonitor(multipleFileDownload, downloads));
final CountDownLatch latch = new CountDownLatch(1);
MultipleFileTransferStateChangeListener transferListener =
new MultipleFileTransferStateChangeListener(latch, multipleFileDownload);
if (objectSummaries.isEmpty()) {
multipleFileDownload.setState(TransferState.Completed);
return multipleFileDownload;
}
for ( S3ObjectSummary summary : objectSummaries ) {
// TODO: non-standard delimiters
File f = new File(destinationDirectory, summary.getKey());
File parentFile = f.getParentFile();
if ( !parentFile.exists() && !parentFile.mkdirs() ) {
throw new RuntimeException("Couldn't create parent directories for " + f.getAbsolutePath());
}
// All the single-file downloads share the same
// MultipleFileTransferProgressUpdatingListener and
// MultipleFileTransferStateChangeListener
GetObjectRequest req = new GetObjectRequest(summary.getBucketName(), summary.getKey())
.withGeneralProgressListener(
listener);
PreparedDownloadContext ctx = prepareDownload(req, f, transferListener, null, false, 0,
null, 0L, resumeOnRetry, null);
preparedDownloadContexts.add(ctx);
}
try {
for (PreparedDownloadContext ctx : preparedDownloadContexts) {
downloads.add(submitDownload(ctx));
}
} catch (Throwable t) {
for (DownloadImpl d : downloads) {
try {
d.getMonitor().getFuture().cancel(true);
} catch (Throwable cancelErr) {
log.warn("DownloadImpl could not be aborted", cancelErr);
}
}
throw new SdkClientException(t);
}
// Notify all state changes waiting for the downloads to all be queued
// to wake up and continue.
latch.countDown();
return multipleFileDownload;
}
private boolean leavesRoot(File localBaseDirectory, String key) {
try {
Path targetPath = new File(localBaseDirectory, key).getCanonicalFile().toPath();
Path rootPath = localBaseDirectory.getCanonicalFile().toPath();
return !targetPath.startsWith(rootPath);
} catch (IOException e) {
throw new RuntimeException("Unable to canonicalize paths", e);
}
}
/**
* Uploads all files in the directory given to the bucket named, optionally
* recursing for all subdirectories.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The directory to upload.
* @param includeSubdirectories
* Whether to include subdirectories in the upload. If true,
* files found in subdirectories will be included with an
* appropriate concatenation to the key prefix.
*/
public MultipleFileUpload uploadDirectory(String bucketName, String virtualDirectoryKeyPrefix, File directory, boolean includeSubdirectories) {
return uploadDirectory(bucketName, virtualDirectoryKeyPrefix, directory, includeSubdirectories, null);
}
/**
* Uploads all files in the directory given to the bucket named, optionally
* recursing for all subdirectories.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The directory to upload.
* @param includeSubdirectories
* Whether to include subdirectories in the upload. If true,
* files found in subdirectories will be included with an
* appropriate concatenation to the key prefix.
* @param metadataProvider
* A callback of type ObjectMetadataProvider
which
* is used to provide metadata for each file being uploaded.
*/
public MultipleFileUpload uploadDirectory(String bucketName, String virtualDirectoryKeyPrefix, File directory, boolean includeSubdirectories, ObjectMetadataProvider metadataProvider) {
return uploadDirectory( bucketName, virtualDirectoryKeyPrefix, directory, includeSubdirectories, metadataProvider, null);
}
/**
* Uploads all files in the directory given to the bucket named, optionally
* recursing for all subdirectories.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The directory to upload.
* @param includeSubdirectories
* Whether to include subdirectories in the upload. If true,
* files found in subdirectories will be included with an
* appropriate concatenation to the key prefix.
* @param metadataProvider
* A callback of type ObjectMetadataProvider
which
* is used to provide metadata for each file being uploaded.
* @param taggingProvider
* A callback of type ObjectTaggingProvider
which
* is used to provide the tags for each file being uploaded.
* @return
*/
public MultipleFileUpload uploadDirectory(String bucketName, String virtualDirectoryKeyPrefix, File directory, boolean includeSubdirectories, ObjectMetadataProvider metadataProvider, ObjectTaggingProvider taggingProvider) {
return uploadDirectory(bucketName, virtualDirectoryKeyPrefix, directory, includeSubdirectories, metadataProvider, taggingProvider, null);
}
/**
* Uploads all files in the directory given to the bucket named, optionally
* recursing for all subdirectories.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The directory to upload.
* @param includeSubdirectories
* Whether to include subdirectories in the upload. If true,
* files found in subdirectories will be included with an
* appropriate concatenation to the key prefix.
* @param metadataProvider
* A callback of type ObjectMetadataProvider
which
* is used to provide metadata for each file being uploaded.
* @param taggingProvider
* A callback of type ObjectTaggingProvider
which
* is used to provide the tags for each file being uploaded.
* @param cannedAclProvider
* A callback of type ObjectCannedAclProvider
which
* is used to provide the CannedAccessControlList for each file
* being uploaded.
* @return
*/
public MultipleFileUpload uploadDirectory(String bucketName, String virtualDirectoryKeyPrefix, File directory, boolean includeSubdirectories, ObjectMetadataProvider metadataProvider, ObjectTaggingProvider taggingProvider, ObjectCannedAclProvider cannedAclProvider) {
if ( directory == null || !directory.exists() || !directory.isDirectory() ) {
throw new IllegalArgumentException("Must provide a directory to upload");
}
assertNotObjectLambdaArn(bucketName, "uploadDirectory");
List files = new LinkedList();
listFiles(directory, files, includeSubdirectories);
return uploadFileList(bucketName, virtualDirectoryKeyPrefix, directory, files, metadataProvider, taggingProvider, cannedAclProvider);
}
/**
* Uploads all specified files to the bucket named, constructing
* relative keys depending on the commonParentDirectory given.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The common parent directory of files to upload. The keys
* of the files in the list of files are constructed relative to
* this directory and the virtualDirectoryKeyPrefix.
* @param files
* A list of files to upload. The keys of the files are
* calculated relative to the common parent directory and the
* virtualDirectoryKeyPrefix.
*/
public MultipleFileUpload uploadFileList(String bucketName, String virtualDirectoryKeyPrefix, File directory, List files) {
return uploadFileList(bucketName, virtualDirectoryKeyPrefix, directory, files, null);
}
/**
* Uploads all specified files to the bucket named, constructing
* relative keys depending on the commonParentDirectory given.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The common parent directory of files to upload. The keys
* of the files in the list of files are constructed relative to
* this directory and the virtualDirectoryKeyPrefix.
* @param files
* A list of files to upload. The keys of the files are
* calculated relative to the common parent directory and the
* virtualDirectoryKeyPrefix.
* @param metadataProvider
* A callback of type ObjectMetadataProvider
which
* is used to provide metadata for each file being uploaded.
*/
public MultipleFileUpload uploadFileList(String bucketName, String virtualDirectoryKeyPrefix, File directory, List files,ObjectMetadataProvider metadataProvider) {
return uploadFileList(bucketName, virtualDirectoryKeyPrefix, directory, files, metadataProvider,null);
}
/**
* Uploads all specified files to the bucket named, constructing
* relative keys depending on the commonParentDirectory given.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The common parent directory of files to upload. The keys
* of the files in the list of files are constructed relative to
* this directory and the virtualDirectoryKeyPrefix.
* @param files
* A list of files to upload. The keys of the files are
* calculated relative to the common parent directory and the
* virtualDirectoryKeyPrefix.
* @param metadataProvider
* A callback of type ObjectMetadataProvider
which
* is used to provide metadata for each file being uploaded.
* @param taggingProvider
* A callback of type ObjectTaggingProvider
which
* is used to provide the tags for each file being uploaded.
* @return
*/
public MultipleFileUpload uploadFileList(String bucketName, String virtualDirectoryKeyPrefix, File directory, List files, ObjectMetadataProvider metadataProvider, ObjectTaggingProvider taggingProvider) {
return uploadFileList(bucketName, virtualDirectoryKeyPrefix, directory, files, metadataProvider, taggingProvider, null);
}
/**
* Uploads all specified files to the bucket named, constructing
* relative keys depending on the commonParentDirectory given.
*
* S3 will overwrite any existing objects that happen to have the same key,
* just as when uploading individual files, so use with caution.
*
*
* If you are uploading Amazon Web Services
* KMS-encrypted objects, you need to specify the correct region of the
* bucket on your client and configure Amazon Web Services Signature Version 4 for added
* security. For more information on how to do this, see
* http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#
* specify-signature-version
*
*
* @param bucketName
* The name of the bucket to upload objects to.
* @param virtualDirectoryKeyPrefix
* The key prefix of the virtual directory to upload to. Use the
* null or empty string to upload files to the root of the
* bucket.
* @param directory
* The common parent directory of files to upload. The keys
* of the files in the list of files are constructed relative to
* this directory and the virtualDirectoryKeyPrefix.
* @param files
* A list of files to upload. The keys of the files are
* calculated relative to the common parent directory and the
* virtualDirectoryKeyPrefix.
* @param metadataProvider
* A callback of type ObjectMetadataProvider
which
* is used to provide metadata for each file being uploaded.
* @param taggingProvider
* A callback of type ObjectTaggingProvider
which
* is used to provide the tags for each file being uploaded.
* @param cannedAclProvider
* A callback of type ObjectCannedAclProvider
which
* is used to provide the CannedAccessControlList for each file
* being uploaded.
* @return
*/
public MultipleFileUpload uploadFileList(String bucketName, String virtualDirectoryKeyPrefix, File directory, List files, ObjectMetadataProvider metadataProvider, ObjectTaggingProvider taggingProvider, ObjectCannedAclProvider cannedAclProvider) {
if ( directory == null || !directory.exists() || !directory.isDirectory() ) {
throw new IllegalArgumentException("Must provide a common base directory for uploaded files");
}
assertNotObjectLambdaArn(bucketName, "uploadFileList");
if (virtualDirectoryKeyPrefix == null || virtualDirectoryKeyPrefix.length() == 0) {
virtualDirectoryKeyPrefix = "";
} else if ( !virtualDirectoryKeyPrefix.endsWith("/") ) {
virtualDirectoryKeyPrefix = virtualDirectoryKeyPrefix + "/";
}
/* This is the hook for adding additional progress listeners */
ProgressListenerChain additionalListeners = new ProgressListenerChain();
TransferProgress progress = new TransferProgress();
/*
* Bind additional progress listeners to this
* MultipleFileTransferProgressUpdatingListener to receive
* ByteTransferred events from each single-file upload implementation.
*/
ProgressListener listener = new MultipleFileTransferProgressUpdatingListener(
progress, additionalListeners);
List uploads = new LinkedList();
MultipleFileUploadImpl multipleFileUpload = new MultipleFileUploadImpl("Uploading etc", progress, additionalListeners, virtualDirectoryKeyPrefix, bucketName, uploads);
multipleFileUpload.setMonitor(new MultipleFileTransferMonitor(multipleFileUpload, uploads));
final CountDownLatch latch = new CountDownLatch(1);
MultipleFileTransferStateChangeListener transferListener =
new MultipleFileTransferStateChangeListener(latch, multipleFileUpload);
if (files == null || files.isEmpty()) {
multipleFileUpload.setState(TransferState.Completed);
} else {
/*
* If the absolute path for the common/base directory does NOT end
* in a separator (which is the case for anything but root
* directories), then we know there's still a separator between the
* base directory and the rest of the file's path, so we increment
* the starting position by one.
*/
int startingPosition = directory.getAbsolutePath().length();
if (!(directory.getAbsolutePath().endsWith(File.separator)))
startingPosition++;
long totalSize = 0;
for (File f : files) {
// Check, if file, since only files can be uploaded.
if (f.isFile()) {
totalSize += f.length();
String key = f.getAbsolutePath()
.substring(startingPosition)
.replaceAll("\\\\", "/");
UploadContext uploadContext = new UploadContext(f, bucketName, key);
ObjectMetadata metadata = new ObjectMetadata();
CannedAccessControlList cannedAcl = null;
ObjectTagging objectTagging = null;
// Invoke the callback if it's present.
// The callback allows the user to customize the metadata
// for each file being uploaded.
if (metadataProvider != null) {
metadataProvider.provideObjectMetadata(f, metadata);
}
// Invoke the callback if it's present.
// The callback allows the user to customize the tags
// for each file being uploaded.
if (taggingProvider != null) {
objectTagging = taggingProvider.provideObjectTags(uploadContext);
}
// Invoke the callback if it's present.
// The callback allows the user to customize the canned
// acl for each file being uploaded.
if (cannedAclProvider != null) {
cannedAcl = cannedAclProvider.provideObjectCannedAcl(f);
}
// All the single-file uploads share the same
// MultipleFileTransferProgressUpdatingListener and
// MultipleFileTransferStateChangeListener
uploads.add((UploadImpl) doUpload(
new PutObjectRequest(bucketName,
virtualDirectoryKeyPrefix + key, f)
.withMetadata(metadata)
.withTagging(objectTagging)
.withCannedAcl(cannedAcl)
. withGeneralProgressListener(
listener), transferListener, null, null));
}
}
progress.setTotalBytesToTransfer(totalSize);
}
// Notify all state changes waiting for the uploads to all be queued
// to wake up and continue
latch.countDown();
return multipleFileUpload;
}
/**
* Lists files in the directory given and adds them to the result list
* passed in, optionally adding subdirectories recursively.
*/
private void listFiles(File dir, List results, boolean includeSubDirectories) {
File[] found = dir.listFiles();
if ( found != null ) {
for ( File f : found ) {
if (f.isDirectory()) {
if (includeSubDirectories) {
listFiles(f, results, includeSubDirectories);
}
} else {
results.add(f);
}
}
}
}
/**
*
* Aborts any multipart uploads that were initiated before the specified date.
*
*
* This method is useful for cleaning up any interrupted multipart uploads.
* TransferManager
attempts to abort any failed uploads,
* but in some cases this may not be possible, such as if network connectivity
* is completely lost.
*
*
* @param bucketName
* The name of the bucket containing the multipart uploads to
* abort.
* @param date
* The date indicating which multipart uploads should be aborted.
*/
public void abortMultipartUploads(String bucketName, Date date)
throws AmazonServiceException, AmazonClientException {
assertNotObjectLambdaArn(bucketName, "abortMultipartUploads");
MultipartUploadListing uploadListing = s3.listMultipartUploads(appendSingleObjectUserAgent(
new ListMultipartUploadsRequest(bucketName)));
do {
for (MultipartUpload upload : uploadListing.getMultipartUploads()) {
if (upload.getInitiated().compareTo(date) < 0) {
s3.abortMultipartUpload(appendSingleObjectUserAgent(new AbortMultipartUploadRequest(
bucketName, upload.getKey(), upload.getUploadId())));
}
}
ListMultipartUploadsRequest request = new ListMultipartUploadsRequest(bucketName)
.withUploadIdMarker(uploadListing.getNextUploadIdMarker())
.withKeyMarker(uploadListing.getNextKeyMarker());
uploadListing = s3.listMultipartUploads(appendSingleObjectUserAgent(request));
} while (uploadListing.isTruncated());
}
/**
* Forcefully shuts down this TransferManager instance - currently executing
* transfers will not be allowed to finish. It also by default shuts down
* the underlying Amazon S3 client.
*
* @see #shutdownNow(boolean)
*/
public void shutdownNow() {
shutdownNow(true);
}
/**
* Forcefully shuts down this TransferManager instance - currently executing
* transfers will not be allowed to finish. Callers should use this method
* when they either:
*
* - have already verified that their transfers have completed by checking
* each transfer's state
*
- need to exit quickly and don't mind stopping transfers before they
* complete.
*
*
* Callers should also remember that uploaded parts from an interrupted
* upload may not always be automatically cleaned up, but callers can use
* {@link #abortMultipartUploads(String, Date)} to clean up any upload
* parts.
*
* @param shutDownS3Client
* Whether to shut down the underlying Amazon S3 client.
*/
public void shutdownNow(boolean shutDownS3Client) {
if (shutDownThreadPools) {
executorService.shutdownNow();
timedThreadPool.shutdownNow();
}
if (shutDownS3Client) {
s3.shutdown();
}
}
/**
* Shutdown without interrupting the threads involved, so that, for example,
* any upload in progress can complete without throwing
* {@link AbortedException}.
*/
private void shutdownThreadPools() {
if (shutDownThreadPools) {
executorService.shutdown();
timedThreadPool.shutdown();
}
}
public static X appendSingleObjectUserAgent(X request) {
request.getRequestClientOptions().appendUserAgent(USER_AGENT);
return request;
}
public static X appendMultipartUserAgent(X request) {
request.getRequestClientOptions().appendUserAgent(USER_AGENT_MULTIPART);
return request;
}
private static final String USER_AGENT = TransferManager.class.getName() + "/" + VersionInfoUtils.getVersion();
private static final String USER_AGENT_MULTIPART = TransferManager.class.getName() + "_multipart/" + VersionInfoUtils.getVersion();
private static final String DEFAULT_DELIMITER = "/";
/**
* There is no need for threads from timedThreadPool if there is no more running threads in current process,
* so we need a daemon thread factory for it.
*/
private static final ThreadFactory daemonThreadFactory = new ThreadFactory() {
final AtomicInteger threadCount = new AtomicInteger( 0 );
public Thread newThread(Runnable r) {
int threadNumber = threadCount.incrementAndGet();
Thread thread = new Thread(r);
thread.setDaemon(true);
thread.setName("S3TransferManagerTimedThread-" + threadNumber);
return thread;
}
};
/**
*
* Schedules a new transfer to copy data from one Amazon S3 location to
* another Amazon S3 location. This method is non-blocking and returns
* immediately (before the copy has finished).
*
*
* TransferManager
doesn't support copying of encrypted objects
* whose encryption materials are stored in an instruction file.
*
*
* Use the returned Copy
object to check if the copy is
* complete.
*
*
* If resources are available, the copy request will begin immediately.
* Otherwise, the copy is scheduled and started as soon as resources become
* available.
*
*
* Note: If the {@link TransferManager} is created with a regional S3 client and
* the source & destination buckets are in different regions, use the
* {@link #copy(CopyObjectRequest, AmazonS3, TransferStateChangeListener)} method.
*
*
* @param sourceBucketName
* The name of the bucket from where the object is to be copied.
* @param sourceKey
* The name of the Amazon S3 object.
* @param destinationBucketName
* The name of the bucket to where the Amazon S3 object has to be
* copied.
* @param destinationKey
* The name of the object in the destination bucket.
*
* @return A new Copy
object to use to check the state of the
* copy request being processed.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*
* @see TransferManager#copy(CopyObjectRequest, AmazonS3, TransferStateChangeListener)
*/
public Copy copy(String sourceBucketName, String sourceKey,
String destinationBucketName, String destinationKey)
throws AmazonServiceException, AmazonClientException {
return copy(new CopyObjectRequest(sourceBucketName, sourceKey,
destinationBucketName, destinationKey));
}
/**
*
* Schedules a new transfer to copy data from one Amazon S3 location to
* another Amazon S3 location. This method is non-blocking and returns
* immediately (i.e. before the copy has finished).
*
*
* TransferManager
doesn't support copying of encrypted objects
* whose encryption materials are stored in an instruction file.
*
*
* Use the returned Copy
object to check if the copy is
* complete.
*
*
* If resources are available, the copy request will begin immediately.
* Otherwise, the copy is scheduled and started as soon as resources become
* available.
*
*
* Note: If the {@link TransferManager} is created with a regional S3 client and
* the source & destination buckets are in different regions, use the
* {@link #copy(CopyObjectRequest, AmazonS3, TransferStateChangeListener)} method.
*
*
* @param copyObjectRequest
* The request containing all the parameters for the copy.
*
* @return A new Copy
object to use to check the state of the
* copy request being processed.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*
* @see TransferManager#copy(CopyObjectRequest, AmazonS3, TransferStateChangeListener)
*/
public Copy copy(final CopyObjectRequest copyObjectRequest){
return copy(copyObjectRequest,null);
}
/**
*
* Schedules a new transfer to copy data from one Amazon S3 location to
* another Amazon S3 location. This method is non-blocking and returns
* immediately (i.e. before the copy has finished).
*
*
* TransferManager
doesn't support copying of encrypted objects
* whose encryption materials are stored in an instruction file.
*
*
* Use the returned Copy
object to check if the copy is
* complete.
*
*
* If resources are available, the copy request will begin immediately.
* Otherwise, the copy is scheduled and started as soon as resources become
* available.
*
*
* Note: If the {@link TransferManager} is created with a regional S3 client and
* the source & destination buckets are in different regions, use the
* {@link #copy(CopyObjectRequest, AmazonS3, TransferStateChangeListener)} method.
*
*
* @param copyObjectRequest
* The request containing all the parameters for the copy.
* @param stateChangeListener
* The transfer state change listener to monitor the copy request
* @return A new Copy
object to use to check the state of the
* copy request being processed.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*
* @see TransferManager#copy(CopyObjectRequest, AmazonS3, TransferStateChangeListener)
*/
public Copy copy(final CopyObjectRequest copyObjectRequest,
final TransferStateChangeListener stateChangeListener) throws
AmazonServiceException,
AmazonClientException {
return copy(copyObjectRequest, s3, stateChangeListener);
}
/**
*
* Schedules a new transfer to copy data from one Amazon S3 location to
* another Amazon S3 location. This method is non-blocking and returns
* immediately (i.e. before the copy has finished).
*
*
* Note: You need to use this method if the {@link TransferManager} is created with
* a regional S3 client and the source & destination buckets are in different regions.
*
*
* TransferManager
doesn't support copying of encrypted objects
* whose encryption materials are stored in an instruction file.
*
*
* Use the returned Copy
object to check if the copy is
* complete.
*
*
* If resources are available, the copy request will begin immediately.
* Otherwise, the copy is scheduled and started as soon as resources become
* available.
*
*
* @param copyObjectRequest The request containing all the parameters for the copy.
* @param srcS3 An AmazonS3 client constructed for the region in which the source
* object's bucket is located.
* @param stateChangeListener The transfer state change listener to monitor the copy request
* @return A new Copy
object to use to check the state of the
* copy request being processed.
* @throws AmazonClientException If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException If any errors occurred in Amazon S3 while processing the
* request.
*/
public Copy copy(final CopyObjectRequest copyObjectRequest, final AmazonS3 srcS3,
final TransferStateChangeListener stateChangeListener) throws
AmazonServiceException,
AmazonClientException {
assertNotObjectLambdaArn(copyObjectRequest.getDestinationBucketName(), "copy");
assertNotObjectLambdaArn(copyObjectRequest.getSourceBucketName(), "copy");
appendSingleObjectUserAgent(copyObjectRequest);
assertParameterNotNull(copyObjectRequest.getSourceBucketName(),
"The source bucket name must be specified when a copy request is initiated.");
assertParameterNotNull(copyObjectRequest.getSourceKey(),
"The source object key must be specified when a copy request is initiated.");
assertParameterNotNull(copyObjectRequest.getDestinationBucketName(),
"The destination bucket name must be specified when a copy request is initiated.");
assertParameterNotNull(copyObjectRequest.getDestinationKey(),
"The destination object key must be specified when a copy request is initiated.");
assertParameterNotNull(srcS3, "The srcS3 parameter is mandatory");
String description =
"Copying object from " + copyObjectRequest.getSourceBucketName() + "/" +
copyObjectRequest.getSourceKey() + " to " +
copyObjectRequest.getDestinationBucketName() + "/" +
copyObjectRequest.getDestinationKey();
GetObjectMetadataRequest getObjectMetadataRequest = new GetObjectMetadataRequest(
copyObjectRequest.getSourceBucketName(), copyObjectRequest.getSourceKey())
.withSSECustomerKey(copyObjectRequest.getSourceSSECustomerKey())
.withRequesterPays(copyObjectRequest.isRequesterPays())
.withVersionId(copyObjectRequest.getSourceVersionId())
.withRequestCredentialsProvider(copyObjectRequest.getRequestCredentialsProvider());
ObjectMetadata metadata = srcS3.getObjectMetadata(getObjectMetadataRequest);
TransferProgress transferProgress = new TransferProgress();
transferProgress.setTotalBytesToTransfer(metadata.getContentLength());
ProgressListenerChain listenerChain = new ProgressListenerChain(
new TransferProgressUpdatingListener(transferProgress));
CopyImpl copy = new CopyImpl(description, transferProgress, listenerChain,
stateChangeListener);
CopyCallable copyCallable = new CopyCallable(this, executorService, copy, copyObjectRequest,
metadata, listenerChain);
CopyMonitor watcher = CopyMonitor
.create(this, copy, executorService, copyCallable, copyObjectRequest,
listenerChain);
copy.setMonitor(watcher);
return copy;
}
/**
* Resumes an upload operation. This upload operation uses the same
* configuration {@link TransferManagerConfiguration} as the original
* upload. Any data already uploaded will be skipped, and only the remaining
* will be uploaded to Amazon S3.
*
* @param persistableUpload
* the upload to resume.
* @return A new Upload
object to use to check the state of the
* upload, listen for progress notifications, and otherwise manage
* the upload.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Upload resumeUpload(PersistableUpload persistableUpload) {
assertParameterNotNull(persistableUpload,
"PauseUpload is mandatory to resume a upload.");
assertNotObjectLambdaArn(persistableUpload.getBucketName(), "resumeUpload");
configuration.setMinimumUploadPartSize(persistableUpload.getPartSize());
configuration.setMultipartUploadThreshold(persistableUpload
.getMutlipartUploadThreshold());
return doUpload(new PutObjectRequest(persistableUpload.getBucketName(),
persistableUpload.getKey(), new File(persistableUpload.getFile())), null, null,
persistableUpload);
}
/**
* Resumes an download operation. This download operation uses the same
* configuration as the original download. Any data already fetched will be
* skipped, and only the remaining data is retrieved from Amazon S3.
*
* @param persistableDownload
* the download to resume.
* @return A new Download
object to use to check the state of
* the download, listen for progress notifications, and otherwise
* manage the download.
*
* @throws AmazonClientException
* If any errors are encountered in the client while making the
* request or handling the response.
* @throws AmazonServiceException
* If any errors occurred in Amazon S3 while processing the
* request.
*/
public Download resumeDownload(PersistableDownload persistableDownload) {
assertParameterNotNull(persistableDownload,
"PausedDownload is mandatory to resume a download.");
assertNotObjectLambdaArn(persistableDownload.getBucketName(), "resumeDownload");
GetObjectRequest request = new GetObjectRequest(
persistableDownload.getBucketName(), persistableDownload.getKey(),
persistableDownload.getVersionId());
if (persistableDownload.getRange() != null
&& persistableDownload.getRange().length == 2) {
long[] range = persistableDownload.getRange();
request.setRange(range[0], range[1]);
}
request.setRequesterPays(persistableDownload.isRequesterPays());
request.setResponseHeaders(persistableDownload.getResponseHeaders());
return doDownload(request, new File(persistableDownload.getFile()), null, null,
APPEND_MODE, 0,
persistableDownload);
}
/**
*
* Asserts that the specified parameter value is not null
and if it is,
* throws an IllegalArgumentException
with the specified error message.
*
*
* @param parameterValue
* The parameter value being checked.
* @param errorMessage
* The error message to include in the IllegalArgumentException
* if the specified parameter is null.
*/
private void assertParameterNotNull(Object parameterValue, String errorMessage) {
if (parameterValue == null) throw new IllegalArgumentException(errorMessage);
}
/**
* Releasing all resources created by TransferManager
before it
* is being garbage collected.
*/
@Override
protected void finalize() throws Throwable {
shutdownThreadPools();
}
/**
* If the client has been marked as immutable then throw an {@link
* UnsupportedOperationException}, otherwise do nothing. Should be called by each mutating
* method.
*/
private void checkMutability() {
if (isImmutable) {
throw new UnsupportedOperationException(
"TransferManager is immutable when created with the builder.");
}
}
private static void assertNotObjectLambdaArn(String arn, String operation) {
if (isObjectLambdaArn(arn)) {
String error = String.format("%s does not support S3 Object Lambda resources", operation);
throw new IllegalArgumentException(error);
}
}
private static void assertNotObjectLambdaUrl(URL url, String operation) {
if (isObjectLambdaHost(url)) {
String error = String.format("%s does not support S3 Object Lambda resources", operation);
throw new IllegalArgumentException(error);
}
}
private static boolean isObjectLambdaArn(String arn) {
if (arn == null) {
return false;
}
return arn.startsWith("arn:") && arn.contains(":s3-object-lambda");
}
private static boolean isObjectLambdaHost(URL url) {
String host = url.getHost();
if (host == null) {
return false;
}
return host.contains(".s3-object-lambda.");
}
}