com.azure.storage.blob.specialized.BlockBlobAsyncClient Maven / Gradle / Ivy
Show all versions of azure-storage-blob Show documentation
// Copyright (c) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License.
package com.azure.storage.blob.specialized;
import com.azure.core.annotation.ReturnType;
import com.azure.core.annotation.ServiceClient;
import com.azure.core.annotation.ServiceMethod;
import com.azure.core.http.HttpPipeline;
import com.azure.core.http.rest.Response;
import com.azure.core.http.rest.SimpleResponse;
import com.azure.core.util.BinaryData;
import com.azure.core.util.Context;
import com.azure.core.util.FluxUtil;
import com.azure.core.util.logging.ClientLogger;
import com.azure.storage.blob.BlobAsyncClient;
import com.azure.storage.blob.BlobServiceVersion;
import com.azure.storage.blob.implementation.models.BlockBlobsCommitBlockListHeaders;
import com.azure.storage.blob.implementation.models.BlockBlobsPutBlobFromUrlHeaders;
import com.azure.storage.blob.implementation.models.BlockBlobsUploadHeaders;
import com.azure.storage.blob.implementation.models.EncryptionScope;
import com.azure.storage.blob.models.AccessTier;
import com.azure.storage.blob.models.BlobHttpHeaders;
import com.azure.storage.blob.models.BlobImmutabilityPolicy;
import com.azure.storage.blob.models.BlobRange;
import com.azure.storage.blob.models.BlobRequestConditions;
import com.azure.storage.blob.models.BlockBlobItem;
import com.azure.storage.blob.models.BlockList;
import com.azure.storage.blob.models.BlockListType;
import com.azure.storage.blob.models.BlockLookupList;
import com.azure.storage.blob.models.CpkInfo;
import com.azure.storage.blob.models.CustomerProvidedKey;
import com.azure.storage.blob.options.BlobUploadFromUrlOptions;
import com.azure.storage.blob.options.BlockBlobCommitBlockListOptions;
import com.azure.storage.blob.options.BlockBlobListBlocksOptions;
import com.azure.storage.blob.options.BlockBlobSimpleUploadOptions;
import com.azure.storage.blob.options.BlockBlobStageBlockFromUrlOptions;
import com.azure.storage.blob.options.BlockBlobStageBlockOptions;
import com.azure.storage.common.Utility;
import com.azure.storage.common.implementation.Constants;
import com.azure.storage.common.implementation.StorageImplUtils;
import reactor.core.publisher.Flux;
import reactor.core.publisher.Mono;
import java.net.MalformedURLException;
import java.net.URL;
import java.nio.ByteBuffer;
import java.util.List;
import java.util.Map;
import java.util.Objects;
import static com.azure.core.util.FluxUtil.monoError;
import static com.azure.core.util.FluxUtil.withContext;
import static com.azure.core.util.tracing.Tracer.AZ_TRACING_NAMESPACE_KEY;
import static com.azure.storage.common.Utility.STORAGE_TRACING_NAMESPACE_VALUE;
/**
* Client to a block blob. It may only be instantiated through a {@link SpecializedBlobClientBuilder} or via the method
* {@link BlobAsyncClient#getBlockBlobAsyncClient()}. This class does not hold any state about a particular blob, but is
* instead a convenient way of sending appropriate requests to the resource on the service.
*
*
* Please refer to the
* Azure Docs for more information.
*
*
* Note this client is an async client that returns reactive responses from Spring Reactor Core project
* (https://projectreactor.io/). Calling the methods in this client will NOT start the actual network
* operation, until {@code .subscribe()} is called on the reactive response. You can simply convert one of these
* responses to a {@link java.util.concurrent.CompletableFuture} object through {@link Mono#toFuture()}.
*/
@ServiceClient(builder = SpecializedBlobClientBuilder.class, isAsync = true)
public final class BlockBlobAsyncClient extends BlobAsyncClientBase {
private static final ClientLogger LOGGER = new ClientLogger(BlockBlobAsyncClient.class);
/**
* Indicates the maximum number of bytes that can be sent in a call to upload.
* @deprecated Use {@link #MAX_STAGE_BLOCK_BYTES_LONG}
*/
@Deprecated
public static final int MAX_UPLOAD_BLOB_BYTES = 256 * Constants.MB;
/**
* Indicates the maximum number of bytes that can be sent in a call to upload.
*/
public static final long MAX_UPLOAD_BLOB_BYTES_LONG = 5000L * Constants.MB;
/**
* Indicates the maximum number of bytes that can be sent in a call to stageBlock.
* @deprecated Use {@link #MAX_STAGE_BLOCK_BYTES_LONG}
*/
@Deprecated
public static final int MAX_STAGE_BLOCK_BYTES = 100 * Constants.MB;
/**
* Indicates the maximum number of bytes that can be sent in a call to stageBlock.
*/
public static final long MAX_STAGE_BLOCK_BYTES_LONG = 4000L * Constants.MB;
/**
* Indicates the maximum number of blocks allowed in a block blob.
*/
public static final int MAX_BLOCKS = 50000;
/**
* Package-private constructor for use by {@link SpecializedBlobClientBuilder}.
*
* @param pipeline The pipeline used to send and receive service requests.
* @param url The endpoint where to send service requests.
* @param serviceVersion The version of the service to receive requests.
* @param accountName The storage account name.
* @param containerName The container name.
* @param blobName The blob name.
* @param snapshot The snapshot identifier for the blob, pass {@code null} to interact with the blob directly.
* @param customerProvidedKey Customer provided key used during encryption of the blob's data on the server, pass
* {@code null} to allow the service to use its own encryption.
* @param encryptionScope Encryption scope used during encryption of the blob's data on the server, pass
* {@code null} to allow the service to use its own encryption.
* @param versionId The version identifier for the blob, pass {@code null} to interact with the latest blob version.
*/
BlockBlobAsyncClient(HttpPipeline pipeline, String url, BlobServiceVersion serviceVersion,
String accountName, String containerName, String blobName, String snapshot, CpkInfo customerProvidedKey,
EncryptionScope encryptionScope, String versionId) {
super(pipeline, url, serviceVersion, accountName, containerName, blobName, snapshot, customerProvidedKey,
encryptionScope, versionId);
}
/**
* Creates a new {@link BlockBlobAsyncClient} with the specified {@code encryptionScope}.
*
* @param encryptionScope the encryption scope for the blob, pass {@code null} to use no encryption scope.
* @return a {@link BlockBlobAsyncClient} with the specified {@code encryptionScope}.
*/
@Override
public BlockBlobAsyncClient getEncryptionScopeAsyncClient(String encryptionScope) {
EncryptionScope finalEncryptionScope = null;
if (encryptionScope != null) {
finalEncryptionScope = new EncryptionScope().setEncryptionScope(encryptionScope);
}
return new BlockBlobAsyncClient(getHttpPipeline(), getAccountUrl(), getServiceVersion(), getAccountName(),
getContainerName(), getBlobName(), getSnapshotId(), getCustomerProvidedKey(), finalEncryptionScope,
getVersionId());
}
/**
* Creates a new {@link BlockBlobAsyncClient} with the specified {@code customerProvidedKey}.
*
* @param customerProvidedKey the {@link CustomerProvidedKey} for the blob,
* pass {@code null} to use no customer provided key.
* @return a {@link BlockBlobAsyncClient} with the specified {@code customerProvidedKey}.
*/
@Override
public BlockBlobAsyncClient getCustomerProvidedKeyAsyncClient(CustomerProvidedKey customerProvidedKey) {
CpkInfo finalCustomerProvidedKey = null;
if (customerProvidedKey != null) {
finalCustomerProvidedKey = new CpkInfo()
.setEncryptionKey(customerProvidedKey.getKey())
.setEncryptionKeySha256(customerProvidedKey.getKeySha256())
.setEncryptionAlgorithm(customerProvidedKey.getEncryptionAlgorithm());
}
return new BlockBlobAsyncClient(getHttpPipeline(), getAccountUrl(), getServiceVersion(), getAccountName(),
getContainerName(), getBlobName(), getSnapshotId(), finalCustomerProvidedKey, encryptionScope,
getVersionId());
}
/**
* Creates a new block blob. By default, this method will not overwrite an existing blob. Updating an existing block
* blob overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content
* of the existing blob is overwritten with the new content. To perform a partial update of a block blob's, use
* PutBlock and PutBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
*
Code Samples
*
*
*
* client.upload(data, length).subscribe(response ->
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getContentMd5())));
*
*
*
* @param data The data to write to the blob. Note that this {@code Flux} must be replayable if retries are enabled
* (the default). In other words, the Flux must produce the same data each time it is subscribed to.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data emitted by the {@code Flux}.
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono upload(Flux data, long length) {
return upload(data, length, false);
}
/**
* Creates a new block blob. By default, this method will not overwrite an existing blob. Updating an existing block
* blob overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content
* of the existing blob is overwritten with the new content. To perform a partial update of a block blob's, use
* PutBlock and PutBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
*
Code Samples
*
*
*
* BinaryData.fromFlux(data, length, false)
* .flatMap(binaryData -> client.upload(binaryData))
* .subscribe(response ->
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getContentMd5())));
*
*
*
* @param data The data to write to the block. Note that this {@code BinaryData} must have defined length
* and must be replayable if retries are enabled (the default), see {@link BinaryData#isReplayable()}.
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono upload(BinaryData data) {
return upload(data, false);
}
/**
* Creates a new block blob, or updates the content of an existing block blob. Updating an existing block blob
* overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content of the
* existing blob is overwritten with the new content. To perform a partial update of a block blob's, use PutBlock
* and PutBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
*
Code Samples
*
*
*
* boolean overwrite = false; // Default behavior
* client.upload(data, length, overwrite).subscribe(response ->
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getContentMd5())));
*
*
*
* @param data The data to write to the blob. Note that this {@code Flux} must be replayable if retries are enabled
* (the default). In other words, the Flux must produce the same data each time it is subscribed to.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data emitted by the {@code Flux}.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono upload(Flux data, long length, boolean overwrite) {
BlobRequestConditions blobRequestConditions = new BlobRequestConditions();
if (!overwrite) {
blobRequestConditions.setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return uploadWithResponse(data, length, null, null, null, null, blobRequestConditions)
.flatMap(FluxUtil::toMono);
}
/**
* Creates a new block blob, or updates the content of an existing block blob. Updating an existing block blob
* overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content of the
* existing blob is overwritten with the new content. To perform a partial update of a block blob's, use PutBlock
* and PutBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
*
Code Samples
*
*
*
* boolean overwrite = false; // Default behavior
* BinaryData.fromFlux(data, length, false)
* .flatMap(binaryData -> client.upload(binaryData, overwrite))
* .subscribe(response ->
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getContentMd5())));
*
*
*
* @param data The data to write to the block. Note that this {@code BinaryData} must have defined length
* and must be replayable if retries are enabled (the default), see {@link BinaryData#isReplayable()}.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono upload(BinaryData data, boolean overwrite) {
BlobRequestConditions blobRequestConditions = new BlobRequestConditions();
if (!overwrite) {
blobRequestConditions.setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return uploadWithResponse(
new BlockBlobSimpleUploadOptions(data)
.setRequestConditions(blobRequestConditions))
.flatMap(FluxUtil::toMono);
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlob; the content of the existing blob is overwritten with the new content. To perform a partial update
* of a block blob's, use PutBlock and PutBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* byte[] md5 = MessageDigest.getInstance("MD5").digest("data".getBytes(StandardCharsets.UTF_8));
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
*
* client.uploadWithResponse(data, length, headers, metadata, AccessTier.HOT, md5, requestConditions)
* .subscribe(response -> System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getValue().getContentMd5())));
*
*
*
* @param data The data to write to the blob. Note that this {@code Flux} must be replayable if retries are enabled
* (the default). In other words, the Flux must produce the same data each time it is subscribed to.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data emitted by the {@code Flux}.
* @param headers {@link BlobHttpHeaders}
* @param metadata Metadata to associate with the blob. If there is leading or trailing whitespace in any
* metadata key or value, it must be removed or encoded.
* @param tier {@link AccessTier} for the destination blob.
* @param contentMd5 An MD5 hash of the blob content. This hash is used to verify the integrity of the blob during
* transport. When this header is specified, the storage service compares the hash of the content that has arrived
* with this header value. Note that this MD5 hash is not stored with the blob. If the two hashes do not match, the
* operation will fail.
* @param requestConditions {@link BlobRequestConditions}
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> uploadWithResponse(Flux data, long length, BlobHttpHeaders headers,
Map metadata, AccessTier tier, byte[] contentMd5, BlobRequestConditions requestConditions) {
try {
return this.uploadWithResponse(new BlockBlobSimpleUploadOptions(data, length).setHeaders(headers)
.setMetadata(metadata).setTier(tier).setContentMd5(contentMd5).setRequestConditions(requestConditions));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlob; the content of the existing blob is overwritten with the new content. To perform a partial update
* of a block blob's, use PutBlock and PutBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* Map<String, String> tags = Collections.singletonMap("tag", "value");
* byte[] md5 = MessageDigest.getInstance("MD5").digest("data".getBytes(StandardCharsets.UTF_8));
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
*
* client.uploadWithResponse(new BlockBlobSimpleUploadOptions(data, length).setHeaders(headers)
* .setMetadata(metadata).setTags(tags).setTier(AccessTier.HOT).setContentMd5(md5)
* .setRequestConditions(requestConditions))
* .subscribe(response -> System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getValue().getContentMd5())));
*
*
*
* @param options {@link BlockBlobSimpleUploadOptions}
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> uploadWithResponse(BlockBlobSimpleUploadOptions options) {
try {
return withContext(context -> uploadWithResponse(options, context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
Mono> uploadWithResponse(BlockBlobSimpleUploadOptions options, Context context) {
StorageImplUtils.assertNotNull("options", options);
Mono dataMono;
BinaryData binaryData = options.getData();
if (binaryData == null) {
Flux dataFlux = options.getDataFlux() == null ? Utility.convertStreamToByteBuffer(
options.getDataStream(), options.getLength(), BlobAsyncClient.BLOB_DEFAULT_UPLOAD_BLOCK_SIZE, true)
: options.getDataFlux();
dataMono = BinaryData.fromFlux(dataFlux, options.getLength(), false);
} else {
dataMono = Mono.just(binaryData);
}
BlobRequestConditions requestConditions = options.getRequestConditions() == null ? new BlobRequestConditions()
: options.getRequestConditions();
Context finalContext = context == null ? Context.NONE : context;
BlobImmutabilityPolicy immutabilityPolicy = options.getImmutabilityPolicy() == null
? new BlobImmutabilityPolicy() : options.getImmutabilityPolicy();
return dataMono.flatMap(data ->
this.azureBlobStorage.getBlockBlobs().uploadWithResponseAsync(containerName, blobName,
options.getLength(), data, null, options.getContentMd5(), options.getMetadata(),
requestConditions.getLeaseId(), options.getTier(), requestConditions.getIfModifiedSince(),
requestConditions.getIfUnmodifiedSince(), requestConditions.getIfMatch(),
requestConditions.getIfNoneMatch(), requestConditions.getTagsConditions(), null,
tagsToString(options.getTags()), immutabilityPolicy.getExpiryTime(), immutabilityPolicy.getPolicyMode(),
options.isLegalHold(), null, options.getHeaders(), getCustomerProvidedKey(),
encryptionScope, finalContext.addData(AZ_TRACING_NAMESPACE_KEY, STORAGE_TRACING_NAMESPACE_VALUE))
.map(rb -> {
BlockBlobsUploadHeaders hd = rb.getDeserializedHeaders();
BlockBlobItem item = new BlockBlobItem(hd.getETag(), hd.getLastModified(), hd.getContentMD5(),
hd.isXMsRequestServerEncrypted(), hd.getXMsEncryptionKeySha256(), hd.getXMsEncryptionScope(),
hd.getXMsVersionId());
return new SimpleResponse<>(rb, item);
}));
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlobFromUrl; the content of the existing blob is overwritten with the new content.
* For more information, see the
* Azure Docs.
*
*
Code Samples
*
*
*
* client.uploadFromUrl(sourceUrl)
* .subscribe(response ->
* System.out.printf("Uploaded BlockBlob from URL, MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getContentMd5())));
*
*
*
* @param sourceUrl The source URL to upload from.
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono uploadFromUrl(String sourceUrl) {
return uploadFromUrl(sourceUrl, false);
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlobFromUrl; the content of the existing blob is overwritten with the new content.
* For more information, see the
* Azure Docs.
*
*
Code Samples
*
*
*
* boolean overwrite = false; // Default behavior
* client.uploadFromUrl(sourceUrl, overwrite).subscribe(response ->
* System.out.printf("Uploaded BlockBlob from URL, MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getContentMd5())));
*
*
*
* @param sourceUrl The source URL to upload from.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono uploadFromUrl(String sourceUrl, boolean overwrite) {
BlobRequestConditions blobRequestConditions = new BlobRequestConditions();
if (!overwrite) {
blobRequestConditions.setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
try {
return uploadFromUrlWithResponse(new BlobUploadFromUrlOptions(sourceUrl)
.setDestinationRequestConditions(blobRequestConditions))
.flatMap(FluxUtil::toMono);
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlobFromUrl; the content of the existing blob is overwritten with the new content.
* For more information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* Map<String, String> tags = Collections.singletonMap("tag", "value");
* byte[] md5 = MessageDigest.getInstance("MD5").digest("data".getBytes(StandardCharsets.UTF_8));
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
*
* client.uploadFromUrlWithResponse(new BlobUploadFromUrlOptions(sourceUrl).setHeaders(headers)
* .setTags(tags).setTier(AccessTier.HOT).setContentMd5(md5)
* .setDestinationRequestConditions(requestConditions))
* .subscribe(response -> System.out.printf("Uploaded BlockBlob from URL, MD5 is %s%n",
* Base64.getEncoder().encodeToString(response.getValue().getContentMd5())));
*
*
*
* @param options {@link BlobUploadFromUrlOptions}
* @return A reactive response containing the information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> uploadFromUrlWithResponse(BlobUploadFromUrlOptions options) {
try {
return withContext(context -> uploadFromUrlWithResponse(options, context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
Mono> uploadFromUrlWithResponse(BlobUploadFromUrlOptions options, Context context) {
StorageImplUtils.assertNotNull("options", options);
BlobRequestConditions destinationRequestConditions =
options.getDestinationRequestConditions() == null ? new BlobRequestConditions()
: options.getDestinationRequestConditions();
BlobRequestConditions sourceRequestConditions =
options.getSourceRequestConditions() == null ? new BlobRequestConditions()
: options.getSourceRequestConditions();
context = context == null ? Context.NONE : context;
String sourceAuth = options.getSourceAuthorization() == null
? null : options.getSourceAuthorization().toString();
try {
new URL(options.getSourceUrl());
} catch (MalformedURLException ex) {
throw LOGGER.logExceptionAsError(new IllegalArgumentException("'sourceUrl' is not a valid url.", ex));
}
// TODO (kasobol-msft) add metadata back (https://github.com/Azure/azure-sdk-for-net/issues/15969)
return this.azureBlobStorage.getBlockBlobs().putBlobFromUrlWithResponseAsync(
containerName, blobName, 0, options.getSourceUrl(), null, null, null,
destinationRequestConditions.getLeaseId(), options.getTier(),
destinationRequestConditions.getIfModifiedSince(), destinationRequestConditions.getIfUnmodifiedSince(),
destinationRequestConditions.getIfMatch(), destinationRequestConditions.getIfNoneMatch(),
destinationRequestConditions.getTagsConditions(),
sourceRequestConditions.getIfModifiedSince(), sourceRequestConditions.getIfUnmodifiedSince(),
sourceRequestConditions.getIfMatch(), sourceRequestConditions.getIfNoneMatch(),
sourceRequestConditions.getTagsConditions(),
null, options.getContentMd5(), tagsToString(options.getTags()),
options.isCopySourceBlobProperties(), sourceAuth, options.getCopySourceTagsMode(), options.getHeaders(),
getCustomerProvidedKey(), encryptionScope,
context.addData(AZ_TRACING_NAMESPACE_KEY, STORAGE_TRACING_NAMESPACE_VALUE))
.map(rb -> {
BlockBlobsPutBlobFromUrlHeaders hd = rb.getDeserializedHeaders();
BlockBlobItem item = new BlockBlobItem(hd.getETag(), hd.getLastModified(), hd.getContentMD5(),
hd.isXMsRequestServerEncrypted(), hd.getXMsEncryptionKeySha256(), hd.getXMsEncryptionScope(),
hd.getXMsVersionId());
return new SimpleResponse<>(rb, item);
});
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param data The data to write to the block. Note that this {@code Flux} must be replayable if retries are enabled
* (the default). In other words, the {@code Flux} must produce the same data each time it is subscribed to.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data emitted by the {@code Flux}.
*
* @return A reactive response signalling completion.
*
*
Code Samples
*
*
*
* client.stageBlock(base64BlockID, data, length)
* .subscribe(
* response -> System.out.println("Staging block completed"),
* error -> System.out.printf("Error when calling stage Block: %s", error));
*
*
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono stageBlock(String base64BlockId, Flux data, long length) {
return stageBlockWithResponse(base64BlockId, data, length, null, null).flatMap(FluxUtil::toMono);
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default),
* see {@link BinaryData#isReplayable()}.
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param data The data to write to the block. Note that this {@code BinaryData} must have defined length
* and must be replayable if retries are enabled (the default), see {@link BinaryData#isReplayable()}.
*
* @return A reactive response signalling completion.
*
*
Code Samples
*
*
*
* BinaryData.fromFlux(data, length, false)
* .flatMap(binaryData -> client.stageBlock(base64BlockID, binaryData))
* .subscribe(
* response -> System.out.println("Staging block completed"),
* error -> System.out.printf("Error when calling stage Block: %s", error));
*
*
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono stageBlock(String base64BlockId, BinaryData data) {
return stageBlockWithResponse(new BlockBlobStageBlockOptions(base64BlockId, data)).flatMap(FluxUtil::toMono);
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default). In other words, the
* {@code Flux} must produce the same data each time it is subscribed to.
*
*
Code Samples
*
*
*
* client.stageBlockWithResponse(base64BlockID, data, length, md5, leaseId).subscribe(response ->
* System.out.printf("Staging block completed with status %d%n", response.getStatusCode()));
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param data The data to write to the block. Note that this {@code Flux} must be replayable if retries are enabled
* (the default). In other words, the Flux must produce the same data each time it is subscribed to.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data emitted by the {@code Flux}.
* @param contentMd5 An MD5 hash of the block content. This hash is used to verify the integrity of the block during
* transport. When this header is specified, the storage service compares the hash of the content that has arrived
* with this header value. Note that this MD5 hash is not stored with the blob. If the two hashes do not match, the
* operation will fail.
* @param leaseId The lease ID the active lease on the blob must match.
*
* @return A reactive response signalling completion.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> stageBlockWithResponse(String base64BlockId, Flux data, long length,
byte[] contentMd5, String leaseId) {
try {
return withContext(context -> stageBlockWithResponse(base64BlockId, data, length,
contentMd5, leaseId, context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Note that the data passed must be replayable if retries are enabled (the default),
* see {@link BinaryData#isReplayable()}.
*
*
Code Samples
*
*
*
* BinaryData.fromFlux(data, length, false)
* .flatMap(binaryData -> client.stageBlockWithResponse(
* new BlockBlobStageBlockOptions(base64BlockID, binaryData)
* .setContentMd5(md5)
* .setLeaseId(leaseId)))
* .subscribe(response ->
* System.out.printf("Staging block completed with status %d%n", response.getStatusCode()));
*
*
*
* @param options {@link BlockBlobStageBlockOptions}
*
* @return A reactive response signalling completion.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> stageBlockWithResponse(BlockBlobStageBlockOptions options) {
Objects.requireNonNull(options, "options must not be null");
try {
return withContext(context -> stageBlockWithResponse(
options.getBase64BlockId(), options.getData(),
options.getContentMd5(), options.getLeaseId(), context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
Mono> stageBlockWithResponse(String base64BlockId, Flux data, long length,
byte[] contentMd5, String leaseId, Context context) {
return BinaryData.fromFlux(data, length, false)
.flatMap(
binaryData -> stageBlockWithResponse(base64BlockId, binaryData, contentMd5, leaseId, context));
}
Mono> stageBlockWithResponse(String base64BlockId, BinaryData data,
byte[] contentMd5, String leaseId, Context context) {
Objects.requireNonNull(data, "data must not be null");
Objects.requireNonNull(data.getLength(), "data must have defined length");
context = context == null ? Context.NONE : context;
return this.azureBlobStorage.getBlockBlobs().stageBlockWithResponseAsync(containerName, blobName,
base64BlockId, data.getLength(), data, contentMd5, null, null,
leaseId, null, getCustomerProvidedKey(),
encryptionScope, context.addData(AZ_TRACING_NAMESPACE_KEY, STORAGE_TRACING_NAMESPACE_VALUE))
.map(response -> new SimpleResponse<>(response, null));
}
/**
* Creates a new block to be committed as part of a blob where the contents are read from a URL. For more
* information, see the Azure
* Docs.
*
* Code Samples
*
*
*
* client.stageBlockFromUrl(base64BlockID, sourceUrl, new BlobRange(offset, count))
* .subscribe(
* response -> System.out.println("Staging block completed"),
* error -> System.out.printf("Error when calling stage Block: %s", error));
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param sourceUrl The url to the blob that will be the source of the copy. A source blob in the same storage
* account can be authenticated via Shared Key. However, if the source is a blob in another account, the source blob
* must either be public or must be authenticated via a shared access signature. If the source blob is public, no
* authentication is required to perform the operation.
* @param sourceRange {@link BlobRange}
*
* @return A reactive response signalling completion.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono stageBlockFromUrl(String base64BlockId, String sourceUrl, BlobRange sourceRange) {
return this.stageBlockFromUrlWithResponse(base64BlockId, sourceUrl, sourceRange, null, null, null)
.flatMap(FluxUtil::toMono);
}
/**
* Creates a new block to be committed as part of a blob where the contents are read from a URL. For more
* information, see the Azure
* Docs.
*
* Code Samples
*
*
*
* BlobRequestConditions sourceRequestConditions = new BlobRequestConditions()
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
*
* client.stageBlockFromUrlWithResponse(base64BlockID, sourceUrl, new BlobRange(offset, count), null,
* leaseId, sourceRequestConditions).subscribe(response ->
* System.out.printf("Staging block from URL completed with status %d%n", response.getStatusCode()));
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param sourceUrl The url to the blob that will be the source of the copy. A source blob in the same storage
* account can be authenticated via Shared Key. However, if the source is a blob in another account, the source blob
* must either be public or must be authenticated via a shared access signature. If the source blob is public, no
* authentication is required to perform the operation.
* @param sourceRange {@link BlobRange}
* @param sourceContentMd5 An MD5 hash of the block content. This hash is used to verify the integrity of the block
* during transport. When this header is specified, the storage service compares the hash of the content that has
* arrived with this header value. Note that this MD5 hash is not stored with the blob. If the two hashes do not
* match, the operation will fail.
* @param leaseId The lease ID that the active lease on the blob must match.
* @param sourceRequestConditions {@link BlobRequestConditions}
* @return A reactive response signalling completion.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> stageBlockFromUrlWithResponse(String base64BlockId, String sourceUrl,
BlobRange sourceRange, byte[] sourceContentMd5, String leaseId, BlobRequestConditions sourceRequestConditions) {
return this.stageBlockFromUrlWithResponse(new BlockBlobStageBlockFromUrlOptions(base64BlockId, sourceUrl)
.setSourceRange(sourceRange).setSourceContentMd5(sourceContentMd5).setLeaseId(leaseId)
.setSourceRequestConditions(sourceRequestConditions));
}
/**
* Creates a new block to be committed as part of a blob where the contents are read from a URL. For more
* information, see the Azure
* Docs.
*
* Code Samples
*
*
*
* BlobRequestConditions sourceRequestConditions = new BlobRequestConditions()
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
*
* client.stageBlockFromUrlWithResponse(new BlockBlobStageBlockFromUrlOptions(base64BlockID, sourceUrl)
* .setSourceRange(new BlobRange(offset, count)).setLeaseId(leaseId)
* .setSourceRequestConditions(sourceRequestConditions)).subscribe(response ->
* System.out.printf("Staging block from URL completed with status %d%n", response.getStatusCode()));
*
*
*
* @param options parameters for the operation.
* @return A reactive response signalling completion.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> stageBlockFromUrlWithResponse(BlockBlobStageBlockFromUrlOptions options) {
try {
return withContext(context -> stageBlockFromUrlWithResponse(options, context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
Mono> stageBlockFromUrlWithResponse(BlockBlobStageBlockFromUrlOptions options, Context context) {
BlobRange sourceRange = (options.getSourceRange() == null) ? new BlobRange(0) : options.getSourceRange();
BlobRequestConditions sourceRequestConditions = (options.getSourceRequestConditions() == null)
? new BlobRequestConditions() : options.getSourceRequestConditions();
try {
new URL(options.getSourceUrl());
} catch (MalformedURLException ex) {
throw LOGGER.logExceptionAsError(new IllegalArgumentException("'sourceUrl' is not a valid url.", ex));
}
context = context == null ? Context.NONE : context;
String sourceAuth = options.getSourceAuthorization() == null
? null : options.getSourceAuthorization().toString();
return this.azureBlobStorage.getBlockBlobs().stageBlockFromURLWithResponseAsync(containerName, blobName,
options.getBase64BlockId(), 0, options.getSourceUrl(), sourceRange.toHeaderValue(), options.getSourceContentMd5(), null, null,
options.getLeaseId(), sourceRequestConditions.getIfModifiedSince(),
sourceRequestConditions.getIfUnmodifiedSince(), sourceRequestConditions.getIfMatch(),
sourceRequestConditions.getIfNoneMatch(), null, sourceAuth, getCustomerProvidedKey(),
encryptionScope, context.addData(AZ_TRACING_NAMESPACE_KEY, STORAGE_TRACING_NAMESPACE_VALUE))
.map(response -> new SimpleResponse<>(response, null));
}
/**
* Returns the list of blocks that have been uploaded as part of a block blob using the specified block list filter.
* For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* client.listBlocks(BlockListType.ALL).subscribe(block -> {
* System.out.println("Committed Blocks:");
* block.getCommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
* System.out.println("Uncommitted Blocks:");
* block.getUncommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
* });
*
*
*
* @param listType Specifies which type of blocks to return.
*
* @return A reactive response containing the list of blocks.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono listBlocks(BlockListType listType) {
return this.listBlocksWithResponse(listType, null).map(Response::getValue);
}
/**
* Returns the list of blocks that have been uploaded as part of a block blob using the specified block list
* filter.
* For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* client.listBlocksWithResponse(BlockListType.ALL, leaseId).subscribe(response -> {
* BlockList block = response.getValue();
* System.out.println("Committed Blocks:");
* block.getCommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
* System.out.println("Uncommitted Blocks:");
* block.getUncommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
* });
*
*
*
* @param listType Specifies which type of blocks to return.
* @param leaseId The lease ID the active lease on the blob must match.
* @return A reactive response containing the list of blocks.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> listBlocksWithResponse(BlockListType listType, String leaseId) {
return this.listBlocksWithResponse(new BlockBlobListBlocksOptions(listType).setLeaseId(leaseId));
}
/**
* Returns the list of blocks that have been uploaded as part of a block blob using the specified block list
* filter.
* For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* client.listBlocksWithResponse(new BlockBlobListBlocksOptions(BlockListType.ALL)
* .setLeaseId(leaseId)
* .setIfTagsMatch(tags)).subscribe(response -> {
* BlockList block = response.getValue();
* System.out.println("Committed Blocks:");
* block.getCommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(),
* b.getSizeLong()));
*
* System.out.println("Uncommitted Blocks:");
* block.getUncommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(),
* b.getSizeLong()));
* });
*
*
*
* @param options {@link BlockBlobListBlocksOptions}
* @return A reactive response containing the list of blocks.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> listBlocksWithResponse(BlockBlobListBlocksOptions options) {
try {
return withContext(context -> listBlocksWithResponse(options, context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
Mono> listBlocksWithResponse(BlockBlobListBlocksOptions options, Context context) {
StorageImplUtils.assertNotNull("options", options);
return this.azureBlobStorage.getBlockBlobs().getBlockListWithResponseAsync(
containerName, blobName, options.getType(), getSnapshotId(), null, options.getLeaseId(),
options.getIfTagsMatch(), null, context)
.map(response -> new SimpleResponse<>(response, response.getValue()));
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part of
* a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can call
* commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new and
* existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* client.commitBlockList(Collections.singletonList(base64BlockID)).subscribe(response ->
* System.out.printf("Committing block list completed. Last modified: %s%n", response.getLastModified()));
*
*
*
* @param base64BlockIds A list of base64 encode {@code String}s that specifies the block IDs to be committed.
* @return A reactive response containing the information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono commitBlockList(List base64BlockIds) {
return commitBlockList(base64BlockIds, false);
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part of
* a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can call
* commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new and
* existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* boolean overwrite = false; // Default behavior
* client.commitBlockList(Collections.singletonList(base64BlockID), overwrite).subscribe(response ->
* System.out.printf("Committing block list completed. Last modified: %s%n", response.getLastModified()));
*
*
*
* @param base64BlockIds A list of base64 encode {@code String}s that specifies the block IDs to be committed.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return A reactive response containing the information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono commitBlockList(List base64BlockIds, boolean overwrite) {
BlobRequestConditions requestConditions = null;
if (!overwrite) {
requestConditions = new BlobRequestConditions().setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return commitBlockListWithResponse(base64BlockIds, null, null, null, requestConditions)
.flatMap(FluxUtil::toMono);
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part
* of a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can
* call commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new
* and existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* client.commitBlockListWithResponse(Collections.singletonList(base64BlockID), headers, metadata,
* AccessTier.HOT, requestConditions).subscribe(response ->
* System.out.printf("Committing block list completed with status %d%n", response.getStatusCode()));
*
*
*
* @param base64BlockIds A list of base64 encode {@code String}s that specifies the block IDs to be committed.
* @param headers {@link BlobHttpHeaders}
* @param metadata Metadata to associate with the blob. If there is leading or trailing whitespace in any
* metadata key or value, it must be removed or encoded.
* @param tier {@link AccessTier} for the destination blob.
* @param requestConditions {@link BlobRequestConditions}
* @return A reactive response containing the information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> commitBlockListWithResponse(List base64BlockIds,
BlobHttpHeaders headers, Map metadata, AccessTier tier,
BlobRequestConditions requestConditions) {
return this.commitBlockListWithResponse(new BlockBlobCommitBlockListOptions(base64BlockIds)
.setHeaders(headers).setMetadata(metadata).setTier(tier).setRequestConditions(requestConditions));
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part
* of a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can
* call commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new
* and existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* Map<String, String> tags = Collections.singletonMap("tag", "value");
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* client.commitBlockListWithResponse(new BlockBlobCommitBlockListOptions(Collections.singletonList(base64BlockID))
* .setHeaders(headers).setMetadata(metadata).setTags(tags).setTier(AccessTier.HOT)
* .setRequestConditions(requestConditions))
* .subscribe(response ->
* System.out.printf("Committing block list completed with status %d%n", response.getStatusCode()));
*
*
*
* @param options {@link BlockBlobCommitBlockListOptions}
* @return A reactive response containing the information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Mono> commitBlockListWithResponse(BlockBlobCommitBlockListOptions options) {
try {
return withContext(context -> commitBlockListWithResponse(options, context));
} catch (RuntimeException ex) {
return monoError(LOGGER, ex);
}
}
Mono> commitBlockListWithResponse(BlockBlobCommitBlockListOptions options,
Context context) {
StorageImplUtils.assertNotNull("options", options);
BlobRequestConditions requestConditions = options.getRequestConditions() == null ? new BlobRequestConditions()
: options.getRequestConditions();
context = context == null ? Context.NONE : context;
BlobImmutabilityPolicy immutabilityPolicy = options.getImmutabilityPolicy() == null
? new BlobImmutabilityPolicy() : options.getImmutabilityPolicy();
return this.azureBlobStorage.getBlockBlobs().commitBlockListWithResponseAsync(containerName, blobName,
new BlockLookupList().setLatest(options.getBase64BlockIds()), null, null, null, options.getMetadata(),
requestConditions.getLeaseId(), options.getTier(), requestConditions.getIfModifiedSince(),
requestConditions.getIfUnmodifiedSince(), requestConditions.getIfMatch(),
requestConditions.getIfNoneMatch(), requestConditions.getTagsConditions(), null,
tagsToString(options.getTags()), immutabilityPolicy.getExpiryTime(), immutabilityPolicy.getPolicyMode(),
options.isLegalHold(), options.getHeaders(), getCustomerProvidedKey(),
encryptionScope, context.addData(AZ_TRACING_NAMESPACE_KEY, STORAGE_TRACING_NAMESPACE_VALUE))
.map(rb -> {
BlockBlobsCommitBlockListHeaders hd = rb.getDeserializedHeaders();
BlockBlobItem item = new BlockBlobItem(hd.getETag(), hd.getLastModified(), hd.getContentMD5(),
hd.isXMsRequestServerEncrypted(), hd.getXMsEncryptionKeySha256(), hd.getXMsEncryptionScope(),
hd.getXMsVersionId());
return new SimpleResponse<>(rb, item);
});
}
}