com.amazonaws.mobileconnectors.s3.transfermanager.internal.UploadCallable Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of aws-java-sdk-s3 Show documentation
Show all versions of aws-java-sdk-s3 Show documentation
The AWS Java Mobile SDK for Amazon S3 module holds the client classes that are used for communicating with Amazon Simple Storage Service
/*
* Copyright 2010-2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License").
* You may not use this file except in compliance with the License.
* A copy of the License is located at
*
* http://aws.amazon.com/apache2.0
*
* or in the "license" file accompanying this file. This file is distributed
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
* express or implied. See the License for the specific language governing
* permissions and limitations under the License.
*/
package com.amazonaws.mobileconnectors.s3.transfermanager.internal;
import com.amazonaws.event.ProgressEvent;
import com.amazonaws.event.ProgressListenerCallbackExecutor;
import com.amazonaws.event.ProgressListenerChain;
import com.amazonaws.mobileconnectors.s3.transfermanager.PersistableUpload;
import com.amazonaws.mobileconnectors.s3.transfermanager.Transfer.TransferState;
import com.amazonaws.mobileconnectors.s3.transfermanager.TransferManager;
import com.amazonaws.mobileconnectors.s3.transfermanager.TransferManagerConfiguration;
import com.amazonaws.mobileconnectors.s3.transfermanager.TransferProgress;
import com.amazonaws.mobileconnectors.s3.transfermanager.model.UploadResult;
import com.amazonaws.services.s3.AmazonS3;
import com.amazonaws.services.s3.AmazonS3EncryptionClient;
import com.amazonaws.services.s3.model.AbortMultipartUploadRequest;
import com.amazonaws.services.s3.model.CompleteMultipartUploadRequest;
import com.amazonaws.services.s3.model.CompleteMultipartUploadResult;
import com.amazonaws.services.s3.model.EncryptedInitiateMultipartUploadRequest;
import com.amazonaws.services.s3.model.EncryptedPutObjectRequest;
import com.amazonaws.services.s3.model.InitiateMultipartUploadRequest;
import com.amazonaws.services.s3.model.ListPartsRequest;
import com.amazonaws.services.s3.model.PartETag;
import com.amazonaws.services.s3.model.PartListing;
import com.amazonaws.services.s3.model.PartSummary;
import com.amazonaws.services.s3.model.PutObjectRequest;
import com.amazonaws.services.s3.model.PutObjectResult;
import com.amazonaws.services.s3.model.StorageClass;
import com.amazonaws.services.s3.model.UploadPartRequest;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import java.io.InputStream;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.concurrent.Callable;
import java.util.concurrent.CancellationException;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Future;
public class UploadCallable implements Callable {
private final AmazonS3 s3;
private final ExecutorService threadPool;
private final PutObjectRequest putObjectRequest;
private String multipartUploadId;
private final UploadImpl upload;
private static final Log log = LogFactory.getLog(UploadCallable.class);
private final TransferManagerConfiguration configuration;
private final List> futures = new ArrayList>();
private final ProgressListenerChain listener;
private final TransferProgress transferProgress;
/**
* ETags retrieved from Amazon S3 for a multi-part upload id. These parts
* will be skipped while resuming a paused upload.
*/
private final List eTagsToSkip = new ArrayList();
private PersistableUpload persistableUpload;
public UploadCallable(TransferManager transferManager,
ExecutorService threadPool, UploadImpl upload,
PutObjectRequest putObjectRequest,
ProgressListenerChain progressListenerChain, String uploadId,
TransferProgress transferProgress) {
this.s3 = transferManager.getAmazonS3Client();
this.configuration = transferManager.getConfiguration();
this.threadPool = threadPool;
this.putObjectRequest = putObjectRequest;
this.listener = progressListenerChain;
this.upload = upload;
this.multipartUploadId = uploadId;
this.transferProgress = transferProgress;
}
List> getFutures() {
return futures;
}
/**
* Returns the ETags retrieved from Amazon S3 for a multi-part upload id.
* These parts will be skipped while resuming a paused upload.
*/
List getETags() {
return eTagsToSkip;
}
String getMultipartUploadId() {
return multipartUploadId;
}
/**
* Returns true if this UploadCallable is processing a multipart upload.
*
* @return True if this UploadCallable is processing a multipart upload.
*/
public boolean isMultipartUpload() {
return TransferManagerUtils.shouldUseMultipartUpload(putObjectRequest, configuration);
}
@Override
public UploadResult call() throws Exception {
upload.setState(TransferState.InProgress);
if (isMultipartUpload()) {
fireProgressEvent(ProgressEvent.STARTED_EVENT_CODE);
return uploadInParts();
} else {
return uploadInOneChunk();
}
}
/**
* Uploads the given request in a single chunk and returns the result.
*/
private UploadResult uploadInOneChunk() {
PutObjectResult putObjectResult = s3.putObject(putObjectRequest);
UploadResult uploadResult = new UploadResult();
uploadResult.setBucketName(putObjectRequest.getBucketName());
uploadResult.setKey(putObjectRequest.getKey());
uploadResult.setETag(putObjectResult.getETag());
uploadResult.setVersionId(putObjectResult.getVersionId());
return uploadResult;
}
/**
* Captures the state of the upload.
*/
private void captureUploadStateIfPossible() {
if (putObjectRequest.getSSECustomerKey() == null) {
persistableUpload = new PersistableUpload(putObjectRequest.getBucketName(),
putObjectRequest.getKey(), putObjectRequest.getFile()
.getAbsolutePath(), multipartUploadId,
configuration.getMinimumUploadPartSize(),
configuration.getMultipartUploadThreshold());
notifyPersistableTransferAvailability();
}
}
public PersistableUpload getPersistableUpload() {
return persistableUpload;
}
/**
* Notifies to the callbacks that state is available
*/
private void notifyPersistableTransferAvailability() {
S3ProgressPublisher.publishTransferPersistable(
listener, persistableUpload);
}
/**
* Uploads the request in multiple chunks, submitting each upload chunk task
* to the thread pool and recording its corresponding Future object, as well
* as the multipart upload id.
*/
private UploadResult uploadInParts() throws Exception {
boolean isUsingEncryption = s3 instanceof AmazonS3EncryptionClient;
long optimalPartSize = getOptimalPartSize(isUsingEncryption);
if (multipartUploadId == null) {
multipartUploadId = initiateMultipartUpload(putObjectRequest,
isUsingEncryption);
}
try {
UploadPartRequestFactory requestFactory = new UploadPartRequestFactory(
putObjectRequest, multipartUploadId, optimalPartSize);
if (TransferManagerUtils.isUploadParallelizable(putObjectRequest, isUsingEncryption)) {
captureUploadStateIfPossible();
uploadPartsInParallel(requestFactory, multipartUploadId);
return null;
} else {
return uploadPartsInSeries(requestFactory);
}
} catch (Exception e) {
fireProgressEvent(ProgressEvent.FAILED_EVENT_CODE);
performAbortMultipartUpload();
throw e;
} finally {
if (putObjectRequest.getInputStream() != null) {
try {
putObjectRequest.getInputStream().close();
} catch (Exception e) {
log.warn("Unable to cleanly close input stream: " + e.getMessage(), e);
}
}
}
}
/**
* Performs an
* {@link AmazonS3#abortMultipartUpload(AbortMultipartUploadRequest)}
* operation for the given multi-part upload.
*/
void performAbortMultipartUpload() {
try {
if (multipartUploadId != null)
s3.abortMultipartUpload(new AbortMultipartUploadRequest(
putObjectRequest.getBucketName(), putObjectRequest
.getKey(), multipartUploadId));
} catch (Exception e2) {
log.info(
"Unable to abort multipart upload, you may need to manually remove uploaded parts: "
+ e2.getMessage(), e2);
}
}
/**
* Computes and returns the optimal part size for the upload.
*/
private long getOptimalPartSize(boolean isUsingEncryption) {
long optimalPartSize = TransferManagerUtils.calculateOptimalPartSize(putObjectRequest,
configuration);
if (isUsingEncryption && optimalPartSize % 32 > 0) {
// When using encryption, parts must line up correctly along cipher
// block boundaries
optimalPartSize = optimalPartSize - (optimalPartSize % 32) + 32;
}
log.debug("Calculated optimal part size: " + optimalPartSize);
return optimalPartSize;
}
/**
* Uploads all parts in the request in serial in this thread, then completes
* the upload and returns the result.
*/
private UploadResult uploadPartsInSeries(UploadPartRequestFactory requestFactory) {
final List partETags = new ArrayList();
while (requestFactory.hasMoreRequests()) {
if (threadPool.isShutdown())
throw new CancellationException("TransferManager has been shutdown");
UploadPartRequest uploadPartRequest = requestFactory.getNextUploadPartRequest();
// Mark the stream in case we need to reset it
InputStream inputStream = uploadPartRequest.getInputStream();
if (inputStream != null && inputStream.markSupported()) {
if (uploadPartRequest.getPartSize() >= Integer.MAX_VALUE) {
inputStream.mark(Integer.MAX_VALUE);
} else {
inputStream.mark((int) uploadPartRequest.getPartSize());
}
}
partETags.add(s3.uploadPart(uploadPartRequest).getPartETag());
}
CompleteMultipartUploadResult completeMultipartUploadResult = s3
.completeMultipartUpload(new CompleteMultipartUploadRequest(putObjectRequest
.getBucketName(),
putObjectRequest.getKey(), multipartUploadId, partETags));
UploadResult uploadResult = new UploadResult();
uploadResult.setBucketName(completeMultipartUploadResult.getBucketName());
uploadResult.setKey(completeMultipartUploadResult.getKey());
uploadResult.setETag(completeMultipartUploadResult.getETag());
uploadResult.setVersionId(completeMultipartUploadResult.getVersionId());
return uploadResult;
}
/**
* Submits a callable for each part to upload to our thread pool and records
* its corresponding Future.
*/
private void uploadPartsInParallel(UploadPartRequestFactory requestFactory,
String uploadId) {
Map partNumbers = identifyExistingPartsForResume(uploadId);
while (requestFactory.hasMoreRequests()) {
if (threadPool.isShutdown())
throw new CancellationException("TransferManager has been shutdown");
UploadPartRequest request = requestFactory.getNextUploadPartRequest();
if (partNumbers.containsKey(request.getPartNumber())) {
PartSummary summary = partNumbers.get(request.getPartNumber());
eTagsToSkip.add(new PartETag(request.getPartNumber(), summary
.getETag()));
transferProgress.updateProgress(summary.getSize());
continue;
}
futures.add(threadPool.submit(new UploadPartCallable(s3, request)));
}
}
private Map identifyExistingPartsForResume(
String uploadId) {
Map partNumbers = new HashMap();
if (uploadId == null) {
return partNumbers;
}
int partNumber = 0;
while (true) {
PartListing parts = s3.listParts(new ListPartsRequest(
putObjectRequest.getBucketName(),
putObjectRequest.getKey(), uploadId)
.withPartNumberMarker(partNumber));
for (PartSummary partSummary : parts.getParts()) {
partNumbers.put(partSummary.getPartNumber(), partSummary);
}
if (!parts.isTruncated()) {
return partNumbers;
}
partNumber = parts.getNextPartNumberMarker();
}
}
/**
* Initiates a multipart upload and returns the upload id
*
* @param isUsingEncryption
*/
private String initiateMultipartUpload(PutObjectRequest putObjectRequest,
boolean isUsingEncryption) {
InitiateMultipartUploadRequest initiateMultipartUploadRequest = null;
if (isUsingEncryption && putObjectRequest instanceof EncryptedPutObjectRequest) {
initiateMultipartUploadRequest = new EncryptedInitiateMultipartUploadRequest(
putObjectRequest.getBucketName(), putObjectRequest.getKey()).withCannedACL(
putObjectRequest.getCannedAcl()).withObjectMetadata(
putObjectRequest.getMetadata());
((EncryptedInitiateMultipartUploadRequest) initiateMultipartUploadRequest)
.setMaterialsDescription(((EncryptedPutObjectRequest) putObjectRequest)
.getMaterialsDescription());
} else {
initiateMultipartUploadRequest = new InitiateMultipartUploadRequest(
putObjectRequest.getBucketName(), putObjectRequest.getKey())
.withCannedACL(putObjectRequest.getCannedAcl())
.withObjectMetadata(putObjectRequest.getMetadata());
}
TransferManager.appendMultipartUserAgent(initiateMultipartUploadRequest);
if (putObjectRequest.getStorageClass() != null) {
initiateMultipartUploadRequest.setStorageClass(
StorageClass.fromValue(putObjectRequest.getStorageClass()));
}
if (putObjectRequest.getRedirectLocation() != null) {
initiateMultipartUploadRequest.setRedirectLocation(putObjectRequest
.getRedirectLocation());
}
if (putObjectRequest.getSSECustomerKey() != null) {
initiateMultipartUploadRequest.setSSECustomerKey(putObjectRequest.getSSECustomerKey());
}
String uploadId = s3.initiateMultipartUpload(initiateMultipartUploadRequest).getUploadId();
log.debug("Initiated new multipart upload: " + uploadId);
return uploadId;
}
private void fireProgressEvent(final int eventType) {
ProgressEvent event = new ProgressEvent(0);
event.setEventCode(eventType);
ProgressListenerCallbackExecutor.progressChanged(listener, event);
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy