com.azure.storage.blob.specialized.BlockBlobClient Maven / Gradle / Ivy
Show all versions of azure-storage-blob Show documentation
// Copyright (c) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License.
package com.azure.storage.blob.specialized;
import com.azure.core.annotation.ReturnType;
import com.azure.core.annotation.ServiceClient;
import com.azure.core.annotation.ServiceMethod;
import com.azure.core.exception.UnexpectedLengthException;
import com.azure.core.http.rest.Response;
import com.azure.core.util.BinaryData;
import com.azure.core.util.Context;
import com.azure.core.util.logging.ClientLogger;
import com.azure.storage.blob.BlobAsyncClient;
import com.azure.storage.blob.BlobClient;
import com.azure.storage.blob.BlobClientBuilder;
import com.azure.storage.blob.models.AccessTier;
import com.azure.storage.blob.models.BlobHttpHeaders;
import com.azure.storage.blob.models.BlobRange;
import com.azure.storage.blob.models.BlobRequestConditions;
import com.azure.storage.blob.models.BlobStorageException;
import com.azure.storage.blob.models.BlockBlobItem;
import com.azure.storage.blob.models.BlockList;
import com.azure.storage.blob.models.BlockListType;
import com.azure.storage.blob.models.CpkInfo;
import com.azure.storage.blob.models.CustomerProvidedKey;
import com.azure.storage.blob.models.ParallelTransferOptions;
import com.azure.storage.blob.options.BlobSeekableByteChannelReadOptions;
import com.azure.storage.blob.options.BlobUploadFromUrlOptions;
import com.azure.storage.blob.options.BlockBlobCommitBlockListOptions;
import com.azure.storage.blob.options.BlockBlobListBlocksOptions;
import com.azure.storage.blob.options.BlockBlobOutputStreamOptions;
import com.azure.storage.blob.options.BlockBlobSeekableByteChannelWriteOptions;
import com.azure.storage.blob.options.BlockBlobSimpleUploadOptions;
import com.azure.storage.blob.options.BlockBlobStageBlockFromUrlOptions;
import com.azure.storage.blob.options.BlockBlobStageBlockOptions;
import com.azure.storage.common.Utility;
import com.azure.storage.common.implementation.Constants;
import com.azure.storage.common.implementation.StorageImplUtils;
import com.azure.storage.common.implementation.StorageSeekableByteChannel;
import reactor.core.publisher.Flux;
import reactor.core.publisher.Mono;
import java.io.InputStream;
import java.io.UncheckedIOException;
import java.net.URL;
import java.nio.ByteBuffer;
import java.nio.channels.SeekableByteChannel;
import java.time.Duration;
import java.util.List;
import java.util.Map;
import java.util.Objects;
import static com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout;
/**
* Client to a block blob. It may only be instantiated through a {@link SpecializedBlobClientBuilder} or via the method
* {@link BlobClient#getBlockBlobClient()}. This class does not hold any state about a particular blob, but is instead
* a convenient way of sending appropriate requests to the resource on the service.
*
*
* Please refer to the Azure Docs for more information.
*/
@ServiceClient(builder = SpecializedBlobClientBuilder.class)
public final class BlockBlobClient extends BlobClientBase {
private static final ClientLogger LOGGER = new ClientLogger(BlockBlobClient.class);
private final BlockBlobAsyncClient client;
/**
* Indicates the maximum number of bytes that can be sent in a call to upload.
* @deprecated Use {@link #MAX_STAGE_BLOCK_BYTES_LONG}.
*/
@Deprecated
public static final int MAX_UPLOAD_BLOB_BYTES = BlockBlobAsyncClient.MAX_UPLOAD_BLOB_BYTES;
/**
* Indicates the maximum number of bytes that can be sent in a call to upload.
*/
public static final long MAX_UPLOAD_BLOB_BYTES_LONG = BlockBlobAsyncClient.MAX_UPLOAD_BLOB_BYTES_LONG;
/**
* Indicates the maximum number of bytes that can be sent in a call to stageBlock.
* @deprecated Use {@link #MAX_STAGE_BLOCK_BYTES_LONG}
*/
@Deprecated
public static final int MAX_STAGE_BLOCK_BYTES = BlockBlobAsyncClient.MAX_STAGE_BLOCK_BYTES;
/**
* Indicates the maximum number of bytes that can be sent in a call to stageBlock.
*/
public static final long MAX_STAGE_BLOCK_BYTES_LONG = BlockBlobAsyncClient.MAX_STAGE_BLOCK_BYTES_LONG;
/**
* Indicates the maximum number of blocks allowed in a block blob.
*/
public static final int MAX_BLOCKS = BlockBlobAsyncClient.MAX_BLOCKS;
/**
* Package-private constructor for use by {@link SpecializedBlobClientBuilder}.
*
* @param client the async block blob client
*/
BlockBlobClient(BlockBlobAsyncClient client) {
super(client);
this.client = client;
}
/**
* Creates a new {@link BlockBlobClient} with the specified {@code encryptionScope}.
*
* @param encryptionScope the encryption scope for the blob, pass {@code null} to use no encryption scope.
* @return a {@link BlockBlobClient} with the specified {@code encryptionScope}.
*/
@Override
public BlockBlobClient getEncryptionScopeClient(String encryptionScope) {
return new BlockBlobClient(client.getEncryptionScopeAsyncClient(encryptionScope));
}
/**
* Creates a new {@link BlockBlobClient} with the specified {@code customerProvidedKey}.
*
* @param customerProvidedKey the {@link CustomerProvidedKey} for the blob,
* pass {@code null} to use no customer provided key.
* @return a {@link BlockBlobClient} with the specified {@code customerProvidedKey}.
*/
@Override
public BlockBlobClient getCustomerProvidedKeyClient(CustomerProvidedKey customerProvidedKey) {
return new BlockBlobClient(client.getCustomerProvidedKeyAsyncClient(customerProvidedKey));
}
/**
* Creates and opens an output stream to write data to the block blob.
*
* Note: We recommend you call write with reasonably sized buffers, you can do so by wrapping the BlobOutputStream
* obtained below with a {@link java.io.BufferedOutputStream}.
*
* @return A {@link BlobOutputStream} object used to write data to the blob.
* @throws BlobStorageException If a storage service error occurred.
*/
public BlobOutputStream getBlobOutputStream() {
return getBlobOutputStream(false);
}
/**
* Creates and opens an output stream to write data to the block blob.
*
* Note: We recommend you call write with reasonably sized buffers, you can do so by wrapping the BlobOutputStream
* obtained below with a {@link java.io.BufferedOutputStream}.
*
* @return A {@link BlobOutputStream} object used to write data to the blob.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @throws BlobStorageException If a storage service error occurred.
*/
public BlobOutputStream getBlobOutputStream(boolean overwrite) {
BlobRequestConditions requestConditions = null;
if (!overwrite) {
if (exists()) {
throw LOGGER.logExceptionAsError(new IllegalArgumentException(Constants.BLOB_ALREADY_EXISTS));
}
requestConditions = new BlobRequestConditions().setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return getBlobOutputStream(requestConditions);
}
/**
* Creates and opens an output stream to write data to the block blob. If the blob already exists on the service, it
* will be overwritten.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
* Note: We recommend you call write with reasonably sized buffers, you can do so by wrapping the BlobOutputStream
* obtained below with a {@link java.io.BufferedOutputStream}.
*
* @param requestConditions A {@link BlobRequestConditions} object that represents the access conditions for the
* blob.
*
* @return A {@link BlobOutputStream} object used to write data to the blob.
* @throws BlobStorageException If a storage service error occurred.
*/
public BlobOutputStream getBlobOutputStream(BlobRequestConditions requestConditions) {
return getBlobOutputStream(null, null, null, null, requestConditions);
}
/**
* Creates and opens an output stream to write data to the block blob. If the blob already exists on the service, it
* will be overwritten.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
* Note: We recommend you call write with reasonably sized buffers, you can do so by wrapping the BlobOutputStream
* obtained below with a {@link java.io.BufferedOutputStream}.
*
* @param parallelTransferOptions {@link ParallelTransferOptions} used to configure buffered uploading.
* @param headers {@link BlobHttpHeaders}
* @param metadata Metadata to associate with the blob. If there is leading or trailing whitespace in any
* metadata key or value, it must be removed or encoded.
* @param tier {@link AccessTier} for the destination blob.
* @param requestConditions {@link BlobRequestConditions}
*
* @return A {@link BlobOutputStream} object used to write data to the blob.
* @throws BlobStorageException If a storage service error occurred.
*/
public BlobOutputStream getBlobOutputStream(ParallelTransferOptions parallelTransferOptions,
BlobHttpHeaders headers, Map metadata, AccessTier tier,
BlobRequestConditions requestConditions) {
return this.getBlobOutputStream(new BlockBlobOutputStreamOptions()
.setParallelTransferOptions(parallelTransferOptions).setHeaders(headers).setMetadata(metadata).setTier(tier)
.setRequestConditions(requestConditions));
}
/**
* Creates and opens an output stream to write data to the block blob. If the blob already exists on the service, it
* will be overwritten.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
* Note: We recommend you call write with reasonably sized buffers, you can do so by wrapping the BlobOutputStream
* obtained below with a {@link java.io.BufferedOutputStream}.
*
* @param options {@link BlockBlobOutputStreamOptions}
* @return A {@link BlobOutputStream} object used to write data to the blob.
* @throws BlobStorageException If a storage service error occurred.
*/
public BlobOutputStream getBlobOutputStream(BlockBlobOutputStreamOptions options) {
return getBlobOutputStream(options, null);
}
/**
* Creates and opens an output stream to write data to the block blob. If the blob already exists on the service, it
* will be overwritten.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
* Note: We recommend you call write with reasonably sized buffers, you can do so by wrapping the BlobOutputStream
* obtained below with a {@link java.io.BufferedOutputStream}.
*
* @param options {@link BlockBlobOutputStreamOptions}
* @param context Additional context that is passed through the Http pipeline during the service call.
* @return A {@link BlobOutputStream} object used to write data to the blob.
* @throws BlobStorageException If a storage service error occurred.
*/
public BlobOutputStream getBlobOutputStream(BlockBlobOutputStreamOptions options, Context context) {
BlobAsyncClient blobClient = prepareBuilder().buildAsyncClient();
return BlobOutputStream.blockBlobOutputStream(blobClient, options, context);
}
/**
* Opens a seekable byte channel in write-only mode to upload the blob.
*
* @param options {@link BlobSeekableByteChannelReadOptions}
* @return A SeekableByteChannel
object that represents the channel to use for writing to the blob.
* @throws BlobStorageException If a storage service error occurred.
* @throws NullPointerException if 'options' is null.
*/
public SeekableByteChannel openSeekableByteChannelWrite(BlockBlobSeekableByteChannelWriteOptions options) {
Objects.requireNonNull(options);
// Behavior can support more modes but this client does not currently support them
StorageSeekableByteChannelBlockBlobWriteBehavior.WriteMode internalMode;
long startingPosition = 0L;
if (options.getWriteMode() == BlockBlobSeekableByteChannelWriteOptions.WriteMode.OVERWRITE) {
internalMode = StorageSeekableByteChannelBlockBlobWriteBehavior.WriteMode.OVERWRITE;
} else {
throw LOGGER.logExceptionAsError(new IllegalArgumentException(
"Unsupported value for `options.getWriteMode()`."));
}
return new StorageSeekableByteChannel(
options.getBlockSizeInBytes() != null
? options.getBlockSizeInBytes().intValue()
: BlobAsyncClient.BLOB_DEFAULT_UPLOAD_BLOCK_SIZE,
new StorageSeekableByteChannelBlockBlobWriteBehavior(this, options.getHeaders(), options.getMetadata(),
options.getTags(), options.getTier(), options.getRequestConditions(), internalMode, null),
startingPosition);
}
private BlobClientBuilder prepareBuilder() {
BlobClientBuilder builder = new BlobClientBuilder()
.pipeline(getHttpPipeline())
.endpoint(getBlobUrl())
.snapshot(getSnapshotId())
.serviceVersion(getServiceVersion());
CpkInfo cpk = getCustomerProvidedKey();
if (cpk != null) {
builder.customerProvidedKey(new CustomerProvidedKey(cpk.getEncryptionKey()));
}
return builder;
}
/**
* Creates a new block blob. By default, this method will not overwrite an existing blob. Updating an existing block
* blob overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content
* of the existing blob is overwritten with the new content. To perform a partial update of a block blob's, use
* PutBlock and PutBlockList. For more information, see the
* Azure Docs.
*
*
Code Samples
*
*
*
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(client.upload(data, length).getContentMd5()));
*
*
*
* @param data The data to write to the blob. The data must be markable. This is in order to support retries. If
* the data is not markable, consider using {@link #getBlobOutputStream()} and writing to the returned OutputStream.
* Alternatively, consider wrapping your data source in a {@link java.io.BufferedInputStream} to add mark support.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data provided in the {@link InputStream}.
* @return The information of the uploaded block blob.
* @throws UncheckedIOException If an I/O error occurs
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem upload(InputStream data, long length) {
return upload(data, length, false);
}
/**
* Creates a new block blob. By default, this method will not overwrite an existing blob. Updating an existing block
* blob overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content
* of the existing blob is overwritten with the new content. To perform a partial update of a block blob's, use
* PutBlock and PutBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* BinaryData binaryData = BinaryData.fromStream(data, length);
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(client.upload(binaryData).getContentMd5()));
*
*
*
* @param data The data to write to the block. Note that this {@code BinaryData} must have defined length
* and must be replayable if retries are enabled (the default), see {@link BinaryData#isReplayable()}.
* @return The information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem upload(BinaryData data) {
return upload(data, false);
}
/**
* Creates a new block blob, or updates the content of an existing block blob. Updating an existing block blob
* overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content of the
* existing blob is overwritten with the new content. To perform a partial update of a block blob's, use PutBlock
* and PutBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* boolean overwrite = false;
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(client.upload(data, length, overwrite).getContentMd5()));
*
*
*
* @param data The data to write to the blob. The data must be markable. This is in order to support retries. If
* the data is not markable, consider using {@link #getBlobOutputStream()} and writing to the returned OutputStream.
* Alternatively, consider wrapping your data source in a {@link java.io.BufferedInputStream} to add mark support.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data provided in the {@link InputStream}.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return The information of the uploaded block blob.
* @throws UncheckedIOException If an I/O error occurs
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem upload(InputStream data, long length, boolean overwrite) {
BlobRequestConditions blobRequestConditions = new BlobRequestConditions();
if (!overwrite) {
blobRequestConditions.setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return uploadWithResponse(data, length, null, null, null, null, blobRequestConditions, null, Context.NONE)
.getValue();
}
/**
* Creates a new block blob, or updates the content of an existing block blob. Updating an existing block blob
* overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content of the
* existing blob is overwritten with the new content. To perform a partial update of a block blob's, use PutBlock
* and PutBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* boolean overwrite = false;
* BinaryData binaryData = BinaryData.fromStream(data, length);
* System.out.printf("Uploaded BlockBlob MD5 is %s%n",
* Base64.getEncoder().encodeToString(client.upload(binaryData, overwrite).getContentMd5()));
*
*
*
* @param data The data to write to the block. Note that this {@code BinaryData} must have defined length
* and must be replayable if retries are enabled (the default), see {@link BinaryData#isReplayable()}.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return The information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem upload(BinaryData data, boolean overwrite) {
BlobRequestConditions blobRequestConditions = new BlobRequestConditions();
if (!overwrite) {
blobRequestConditions.setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return uploadWithResponse(
new BlockBlobSimpleUploadOptions(data)
.setRequestConditions(blobRequestConditions),
null, Context.NONE)
.getValue();
}
/**
* Creates a new block blob, or updates the content of an existing block blob. Updating an existing block blob
* overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content of the
* existing blob is overwritten with the new content. To perform a partial update of a block blob's, use PutBlock
* and PutBlockList. For more information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
*
* byte[] md5 = MessageDigest.getInstance("MD5").digest("data".getBytes(StandardCharsets.UTF_8));
*
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Uploaded BlockBlob MD5 is %s%n", Base64.getEncoder()
* .encodeToString(client.uploadWithResponse(data, length, headers, metadata, AccessTier.HOT, md5,
* requestConditions, timeout, context)
* .getValue()
* .getContentMd5()));
*
*
*
* @param data The data to write to the blob. The data must be markable. This is in order to support retries. If
* the data is not markable, consider using {@link #getBlobOutputStream()} and writing to the returned OutputStream.
* Alternatively, consider wrapping your data source in a {@link java.io.BufferedInputStream} to add mark support.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data provided in the {@link InputStream}.
* @param headers {@link BlobHttpHeaders}
* @param metadata Metadata to associate with the blob. If there is leading or trailing whitespace in any
* metadata key or value, it must be removed or encoded.
* @param tier {@link AccessTier} for the destination blob.
* @param contentMd5 An MD5 hash of the block content. This hash is used to verify the integrity of the block during
* transport. When this header is specified, the storage service compares the hash of the content that has arrived
* with this header value. Note that this MD5 hash is not stored with the blob. If the two hashes do not match, the
* operation will fail.
* @param requestConditions {@link BlobRequestConditions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return The information of the uploaded block blob.
*
* @throws UnexpectedLengthException when the length of data does not match the input {@code length}.
* @throws NullPointerException if the input data is null.
* @throws UncheckedIOException If an I/O error occurs
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response uploadWithResponse(InputStream data, long length, BlobHttpHeaders headers,
Map metadata, AccessTier tier, byte[] contentMd5, BlobRequestConditions requestConditions,
Duration timeout, Context context) {
return this.uploadWithResponse(new BlockBlobSimpleUploadOptions(data, length).setHeaders(headers)
.setMetadata(metadata).setTier(tier).setContentMd5(contentMd5).setRequestConditions(requestConditions),
timeout, context);
}
/**
* Creates a new block blob, or updates the content of an existing block blob. Updating an existing block blob
* overwrites any existing metadata on the blob. Partial updates are not supported with PutBlob; the content of the
* existing blob is overwritten with the new content. To perform a partial update of a block blob's, use PutBlock
* and PutBlockList. For more information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* Map<String, String> tags = Collections.singletonMap("tag", "value");
*
* byte[] md5 = MessageDigest.getInstance("MD5").digest("data".getBytes(StandardCharsets.UTF_8));
*
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Uploaded BlockBlob MD5 is %s%n", Base64.getEncoder()
* .encodeToString(client.uploadWithResponse(new BlockBlobSimpleUploadOptions(data, length)
* .setHeaders(headers).setMetadata(metadata).setTags(tags).setTier(AccessTier.HOT).setContentMd5(md5)
* .setRequestConditions(requestConditions), timeout, context)
* .getValue()
* .getContentMd5()));
*
*
*
* @param options {@link BlockBlobSimpleUploadOptions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return The information of the uploaded block blob.
*
* @throws UnexpectedLengthException when the length of data does not match the input {@code length}.
* @throws NullPointerException if the input data is null.
* @throws UncheckedIOException If an I/O error occurs
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response uploadWithResponse(BlockBlobSimpleUploadOptions options, Duration timeout,
Context context) {
StorageImplUtils.assertNotNull("options", options);
Mono> upload = client.uploadWithResponse(options, context);
try {
return blockWithOptionalTimeout(upload, timeout);
} catch (UncheckedIOException e) {
throw LOGGER.logExceptionAsError(e);
}
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlobFromUrl; the content of the existing blob is overwritten with the new content.
* For more information, see the
* Azure Docs.
*
*
Code Samples
*
*
*
* System.out.printf("Uploaded BlockBlob from URL, MD5 is %s%n",
* Base64.getEncoder().encodeToString(client.uploadFromUrl(sourceUrl).getContentMd5()));
*
*
*
* @param sourceUrl The source URL to upload from.
* @return The information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem uploadFromUrl(String sourceUrl) {
return uploadFromUrl(sourceUrl, false);
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlobFromUrl; the content of the existing blob is overwritten with the new content.
* For more information, see the
* Azure Docs.
*
*
Code Samples
*
*
*
* boolean overwrite = false;
* System.out.printf("Uploaded BlockBlob from URL, MD5 is %s%n",
* Base64.getEncoder().encodeToString(client.uploadFromUrl(sourceUrl, overwrite).getContentMd5()));
*
*
*
* @param sourceUrl The source URL to upload from.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return The information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem uploadFromUrl(String sourceUrl, boolean overwrite) {
BlobRequestConditions blobRequestConditions = new BlobRequestConditions();
if (!overwrite) {
blobRequestConditions.setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return uploadFromUrlWithResponse(
new BlobUploadFromUrlOptions(sourceUrl).setDestinationRequestConditions(blobRequestConditions),
null, Context.NONE)
.getValue();
}
/**
* Creates a new block blob, or updates the content of an existing block blob.
*
* Updating an existing block blob overwrites any existing metadata on the blob. Partial updates are not supported
* with PutBlobFromUrl; the content of the existing blob is overwritten with the new content.
* For more information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* Map<String, String> tags = Collections.singletonMap("tag", "value");
*
* byte[] md5 = MessageDigest.getInstance("MD5").digest("data".getBytes(StandardCharsets.UTF_8));
*
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Uploaded BlockBlob MD5 is %s%n", Base64.getEncoder()
* .encodeToString(client.uploadFromUrlWithResponse(new BlobUploadFromUrlOptions(sourceUrl)
* .setHeaders(headers).setTags(tags).setTier(AccessTier.HOT).setContentMd5(md5)
* .setDestinationRequestConditions(requestConditions), timeout, context)
* .getValue()
* .getContentMd5()));
*
*
*
* @param options {@link BlobUploadFromUrlOptions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
* @return The information of the uploaded block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response uploadFromUrlWithResponse(BlobUploadFromUrlOptions options, Duration timeout,
Context context) {
StorageImplUtils.assertNotNull("options", options);
Mono> upload = client.uploadFromUrlWithResponse(options, context);
try {
return blockWithOptionalTimeout(upload, timeout);
} catch (UncheckedIOException e) {
throw LOGGER.logExceptionAsError(e);
}
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* client.stageBlock(base64BlockId, data, length);
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param data The data to write to the block. The data must be markable. This is in order to support retries. If
* the data is not markable, consider using {@link #getBlobOutputStream()} and writing to the returned OutputStream.
* Alternatively, consider wrapping your data source in a {@link java.io.BufferedInputStream} to add mark support.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data provided in the {@link InputStream}.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public void stageBlock(String base64BlockId, InputStream data, long length) {
stageBlockWithResponse(base64BlockId, data, length, null, null, null, Context.NONE);
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* BinaryData binaryData = BinaryData.fromStream(data, length);
* client.stageBlock(base64BlockId, binaryData);
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param data The data to write to the block. Note that this {@code BinaryData} must have defined length
* and must be replayable if retries are enabled (the default), see {@link BinaryData#isReplayable()}.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public void stageBlock(String base64BlockId, BinaryData data) {
stageBlockWithResponse(new BlockBlobStageBlockOptions(base64BlockId, data), null, Context.NONE);
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* Context context = new Context("key", "value");
* System.out.printf("Staging block completed with status %d%n",
* client.stageBlockWithResponse(base64BlockId, data, length, md5, leaseId, timeout, context).getStatusCode());
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param data The data to write to the block. The data must be markable. This is in order to support retries. If
* the data is not markable, consider using {@link #getBlobOutputStream()} and writing to the returned OutputStream.
* Alternatively, consider wrapping your data source in a {@link java.io.BufferedInputStream} to add mark support.
* @param length The exact length of the data. It is important that this value match precisely the length of the
* data provided in the {@link InputStream}.
* @param contentMd5 An MD5 hash of the block content. This hash is used to verify the integrity of the block during
* transport. When this header is specified, the storage service compares the hash of the content that has arrived
* with this header value. Note that this MD5 hash is not stored with the blob. If the two hashes do not match, the
* operation will fail.
* @param leaseId The lease ID the active lease on the blob must match.
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return A response containing status code and HTTP headers
*
* @throws UnexpectedLengthException when the length of data does not match the input {@code length}.
* @throws NullPointerException if the input data is null.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response stageBlockWithResponse(String base64BlockId, InputStream data, long length, byte[] contentMd5,
String leaseId, Duration timeout, Context context) {
StorageImplUtils.assertNotNull("data", data);
Flux fbb = Utility.convertStreamToByteBuffer(data, length,
BlobAsyncClient.BLOB_DEFAULT_UPLOAD_BLOCK_SIZE, true);
Mono> response = client.stageBlockWithResponse(base64BlockId, fbb, length, contentMd5, leaseId,
context);
return blockWithOptionalTimeout(response, timeout);
}
/**
* Uploads the specified block to the block blob's "staging area" to be later committed by a call to
* commitBlockList. For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* Context context = new Context("key", "value");
* BinaryData binaryData = BinaryData.fromStream(data, length);
* BlockBlobStageBlockOptions options = new BlockBlobStageBlockOptions(base64BlockId, binaryData)
* .setContentMd5(md5)
* .setLeaseId(leaseId);
* System.out.printf("Staging block completed with status %d%n",
* client.stageBlockWithResponse(options, timeout, context).getStatusCode());
*
*
*
* @param options {@link BlockBlobStageBlockOptions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return A response containing status code and HTTP headers
*
* @throws UnexpectedLengthException when the length of data does not match the input {@code length}.
* @throws NullPointerException if the input options is null.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response stageBlockWithResponse(BlockBlobStageBlockOptions options, Duration timeout, Context context) {
Objects.requireNonNull(options, "options must not be null");
Mono> response = client.stageBlockWithResponse(
options.getBase64BlockId(), options.getData(), options.getContentMd5(), options.getLeaseId(), context);
return blockWithOptionalTimeout(response, timeout);
}
/**
* Creates a new block to be committed as part of a blob where the contents are read from a URL. For more
* information, see the Azure
* Docs.
*
* Code Samples
*
*
*
* client.stageBlockFromUrl(base64BlockId, sourceUrl, new BlobRange(offset, count));
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param sourceUrl The url to the blob that will be the source of the copy. A source blob in the same storage
* account can be authenticated via Shared Key. However, if the source is a blob in another account, the source blob
* must either be public or must be authenticated via a shared access signature. If the source blob is public, no
* authentication is required to perform the operation.
* @param sourceRange {@link BlobRange}
* @throws IllegalArgumentException If {@code sourceUrl} is a malformed {@link URL}.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public void stageBlockFromUrl(String base64BlockId, String sourceUrl, BlobRange sourceRange) {
stageBlockFromUrlWithResponse(base64BlockId, sourceUrl, sourceRange, null, null, null, null, Context.NONE);
}
/**
* Creates a new block to be committed as part of a blob where the contents are read from a URL. For more
* information, see the Azure
* Docs.
*
* Code Samples
*
*
*
* BlobRequestConditions sourceRequestConditions = new BlobRequestConditions()
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Staging block from URL completed with status %d%n",
* client.stageBlockFromUrlWithResponse(base64BlockId, sourceUrl, new BlobRange(offset, count), null,
* leaseId, sourceRequestConditions, timeout, context).getStatusCode());
*
*
*
* @param base64BlockId A Base64 encoded {@code String} that specifies the ID for this block. Note that all block
* ids for a given blob must be the same length.
* @param sourceUrl The url to the blob that will be the source of the copy. A source blob in the same storage
* account can be authenticated via Shared Key. However, if the source is a blob in another account, the source blob
* must either be public or must be authenticated via a shared access signature. If the source blob is public, no
* authentication is required to perform the operation.
* @param sourceRange {@link BlobRange}
* @param sourceContentMd5 An MD5 hash of the block content. This hash is used to verify the integrity of the block
* during transport. When this header is specified, the storage service compares the hash of the content that has
* arrived with this header value. Note that this MD5 hash is not stored with the blob. If the two hashes do not
* match, the operation will fail.
* @param leaseId The lease ID that the active lease on the blob must match.
* @param sourceRequestConditions {@link BlobRequestConditions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return A response containing status code and HTTP headers
* @throws IllegalArgumentException If {@code sourceUrl} is a malformed {@link URL}.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response stageBlockFromUrlWithResponse(String base64BlockId, String sourceUrl, BlobRange sourceRange,
byte[] sourceContentMd5, String leaseId, BlobRequestConditions sourceRequestConditions, Duration timeout,
Context context) {
return stageBlockFromUrlWithResponse(new BlockBlobStageBlockFromUrlOptions(base64BlockId, sourceUrl)
.setSourceRange(sourceRange).setSourceContentMd5(sourceContentMd5).setLeaseId(leaseId)
.setSourceRequestConditions(sourceRequestConditions), timeout, context);
}
/**
* Creates a new block to be committed as part of a blob where the contents are read from a URL. For more
* information, see the Azure
* Docs.
*
* Code Samples
*
*
*
* BlobRequestConditions sourceRequestConditions = new BlobRequestConditions()
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Staging block from URL completed with status %d%n",
* client.stageBlockFromUrlWithResponse(new BlockBlobStageBlockFromUrlOptions(base64BlockId, sourceUrl)
* .setSourceRange(new BlobRange(offset, count)).setLeaseId(leaseId)
* .setSourceRequestConditions(sourceRequestConditions), timeout, context).getStatusCode());
*
*
*
* @param options Parameters for the operation
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return A response containing status code and HTTP headers
* @throws IllegalArgumentException If {@code sourceUrl} is a malformed {@link URL}.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response stageBlockFromUrlWithResponse(BlockBlobStageBlockFromUrlOptions options, Duration timeout,
Context context) {
Mono> response = client.stageBlockFromUrlWithResponse(options, context);
return blockWithOptionalTimeout(response, timeout);
}
/**
* Returns the list of blocks that have been uploaded as part of a block blob using the specified block list filter.
* For more information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* BlockList block = client.listBlocks(BlockListType.ALL);
*
* System.out.println("Committed Blocks:");
* block.getCommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
* System.out.println("Uncommitted Blocks:");
* block.getUncommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
*
*
* @param listType Specifies which type of blocks to return.
*
* @return The list of blocks.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockList listBlocks(BlockListType listType) {
return this.listBlocksWithResponse(listType, null, null, Context.NONE).getValue();
}
/**
* Returns the list of blocks that have been uploaded as part of a block blob using the specified block list
* filter. For more information, see the Azure Docs.
*
* Code Samples
*
*
*
* Context context = new Context("key", "value");
* BlockList block = client.listBlocksWithResponse(BlockListType.ALL, leaseId, timeout, context).getValue();
*
* System.out.println("Committed Blocks:");
* block.getCommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
* System.out.println("Uncommitted Blocks:");
* block.getUncommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
*
*
* @param listType Specifies which type of blocks to return.
* @param leaseId The lease ID the active lease on the blob must match.
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return The list of blocks.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response listBlocksWithResponse(BlockListType listType, String leaseId, Duration timeout,
Context context) {
return listBlocksWithResponse(new BlockBlobListBlocksOptions(listType).setLeaseId(leaseId), timeout, context);
}
/**
* Returns the list of blocks that have been uploaded as part of a block blob using the specified block list
* filter. For more information, see the Azure Docs.
*
* Code Samples
*
*
*
* Context context = new Context("key", "value");
* BlockList block = client.listBlocksWithResponse(new BlockBlobListBlocksOptions(BlockListType.ALL)
* .setLeaseId(leaseId)
* .setIfTagsMatch(tags), timeout, context).getValue();
*
* System.out.println("Committed Blocks:");
* block.getCommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
* System.out.println("Uncommitted Blocks:");
* block.getUncommittedBlocks().forEach(b -> System.out.printf("Name: %s, Size: %d", b.getName(), b.getSizeLong()));
*
*
*
* @param options {@link BlockBlobListBlocksOptions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return The list of blocks.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response listBlocksWithResponse(BlockBlobListBlocksOptions options, Duration timeout,
Context context) {
return blockWithOptionalTimeout(client.listBlocksWithResponse(options, context), timeout);
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part of
* a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can call
* commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new and
* existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* System.out.printf("Committing block list completed. Last modified: %s%n",
* client.commitBlockList(Collections.singletonList(base64BlockId)).getLastModified());
*
*
*
* @param base64BlockIds A list of base64 encode {@code String}s that specifies the block IDs to be committed.
* @return The information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem commitBlockList(List base64BlockIds) {
return commitBlockList(base64BlockIds, false);
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part of
* a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can call
* commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new and
* existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* Code Samples
*
*
*
* boolean overwrite = false; // Default behavior
* System.out.printf("Committing block list completed. Last modified: %s%n",
* client.commitBlockList(Collections.singletonList(base64BlockId), overwrite).getLastModified());
*
*
*
* @param base64BlockIds A list of base64 encode {@code String}s that specifies the block IDs to be committed.
* @param overwrite Whether to overwrite, should data exist on the blob.
* @return The information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public BlockBlobItem commitBlockList(List base64BlockIds, boolean overwrite) {
BlobRequestConditions requestConditions = null;
if (!overwrite) {
requestConditions = new BlobRequestConditions().setIfNoneMatch(Constants.HeaderConstants.ETAG_WILDCARD);
}
return commitBlockListWithResponse(base64BlockIds, null, null, null, requestConditions, null, Context.NONE)
.getValue();
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part
* of a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can
* call commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new
* and existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Committing block list completed with status %d%n",
* client.commitBlockListWithResponse(Collections.singletonList(base64BlockId), headers, metadata,
* AccessTier.HOT, requestConditions, timeout, context).getStatusCode());
*
*
*
* @param base64BlockIds A list of base64 encode {@code String}s that specifies the block IDs to be committed.
* @param headers {@link BlobHttpHeaders}
* @param metadata Metadata to associate with the blob. If there is leading or trailing whitespace in any
* metadata key or value, it must be removed or encoded.
* @param tier {@link AccessTier} for the destination blob.
* @param requestConditions {@link BlobRequestConditions}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return The information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response commitBlockListWithResponse(List base64BlockIds, BlobHttpHeaders headers,
Map metadata, AccessTier tier, BlobRequestConditions requestConditions, Duration timeout,
Context context) {
return this.commitBlockListWithResponse(new BlockBlobCommitBlockListOptions(base64BlockIds)
.setHeaders(headers).setMetadata(metadata).setTier(tier).setRequestConditions(requestConditions),
timeout, context);
}
/**
* Writes a blob by specifying the list of block IDs that are to make up the blob. In order to be written as part
* of a blob, a block must have been successfully written to the server in a prior stageBlock operation. You can
* call commitBlockList to update a blob by uploading only those blocks that have changed, then committing the new
* and existing blocks together. Any blocks not specified in the block list and permanently deleted. For more
* information, see the
* Azure Docs.
*
* To avoid overwriting, pass "*" to {@link BlobRequestConditions#setIfNoneMatch(String)}.
*
*
Code Samples
*
*
*
* BlobHttpHeaders headers = new BlobHttpHeaders()
* .setContentMd5("data".getBytes(StandardCharsets.UTF_8))
* .setContentLanguage("en-US")
* .setContentType("binary");
*
* Map<String, String> metadata = Collections.singletonMap("metadata", "value");
* Map<String, String> tags = Collections.singletonMap("tag", "value");
* BlobRequestConditions requestConditions = new BlobRequestConditions()
* .setLeaseId(leaseId)
* .setIfUnmodifiedSince(OffsetDateTime.now().minusDays(3));
* Context context = new Context("key", "value");
*
* System.out.printf("Committing block list completed with status %d%n",
* client.commitBlockListWithResponse(
* new BlockBlobCommitBlockListOptions(Collections.singletonList(base64BlockId)).setHeaders(headers)
* .setMetadata(metadata).setTags(tags).setTier(AccessTier.HOT)
* .setRequestConditions(requestConditions), timeout, context)
* .getStatusCode());
*
*
*
* @param options {@link BlockBlobCommitBlockListOptions options}
* @param timeout An optional timeout value beyond which a {@link RuntimeException} will be raised.
* @param context Additional context that is passed through the Http pipeline during the service call.
*
* @return The information of the block blob.
*/
@ServiceMethod(returns = ReturnType.SINGLE)
public Response commitBlockListWithResponse(BlockBlobCommitBlockListOptions options,
Duration timeout, Context context) {
Mono> response = client.commitBlockListWithResponse(
options, context);
return blockWithOptionalTimeout(response, timeout);
}
}