org.apache.cassandra.db.streaming.CassandraStreamWriter Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of cassandra-all Show documentation
Show all versions of cassandra-all Show documentation
The Apache Cassandra Project develops a highly scalable second-generation distributed database, bringing together Dynamo's fully distributed design and Bigtable's ColumnFamily-based data model.
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.cassandra.db.streaming;
import java.io.IOException;
import java.nio.ByteBuffer;
import java.util.Collection;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import net.jpountz.lz4.LZ4Compressor;
import net.jpountz.lz4.LZ4Factory;
import org.apache.cassandra.io.compress.BufferType;
import org.apache.cassandra.io.sstable.format.SSTableFormat.Components;
import org.apache.cassandra.io.sstable.format.SSTableReader;
import org.apache.cassandra.io.util.ChannelProxy;
import org.apache.cassandra.io.util.DataIntegrityMetadata.ChecksumValidator;
import org.apache.cassandra.streaming.ProgressInfo;
import org.apache.cassandra.streaming.StreamManager;
import org.apache.cassandra.streaming.StreamManager.StreamRateLimiter;
import org.apache.cassandra.streaming.StreamSession;
import org.apache.cassandra.streaming.StreamingDataOutputPlus;
import org.apache.cassandra.streaming.async.StreamCompressionSerializer;
import org.apache.cassandra.utils.FBUtilities;
import org.apache.cassandra.utils.memory.BufferPools;
import static org.apache.cassandra.net.MessagingService.current_version;
/**
* CassandraStreamWriter writes given section of the SSTable to given channel.
*/
public class CassandraStreamWriter
{
private static final int DEFAULT_CHUNK_SIZE = 64 * 1024;
private static final Logger logger = LoggerFactory.getLogger(CassandraStreamWriter.class);
protected final SSTableReader sstable;
private final LZ4Compressor compressor = LZ4Factory.fastestInstance().fastCompressor();
protected final Collection sections;
protected final StreamRateLimiter limiter;
protected final StreamSession session;
private final long totalSize;
public CassandraStreamWriter(SSTableReader sstable, CassandraStreamHeader header, StreamSession session)
{
this.session = session;
this.sstable = sstable;
this.sections = header.sections;
this.limiter = StreamManager.getRateLimiter(session.peer);
this.totalSize = header.size();
}
/**
* Stream file of specified sections to given channel.
*
* CassandraStreamWriter uses LZF compression on wire to decrease size to transfer.
*
* @param out where this writes data to
* @throws IOException on any I/O error
*/
public void write(StreamingDataOutputPlus out) throws IOException
{
long totalSize = totalSize();
logger.debug("[Stream #{}] Start streaming file {} to {}, repairedAt = {}, totalSize = {}", session.planId(),
sstable.getFilename(), session.peer, sstable.getSSTableMetadata().repairedAt, totalSize);
try(ChannelProxy proxy = sstable.getDataChannel().newChannel();
ChecksumValidator validator = sstable.maybeGetChecksumValidator())
{
int bufferSize = validator == null ? DEFAULT_CHUNK_SIZE: validator.chunkSize;
// setting up data compression stream
long progress = 0L;
// stream each of the required sections of the file
String filename = sstable.descriptor.fileFor(Components.DATA).toString();
for (SSTableReader.PartitionPositionBounds section : sections)
{
long start = validator == null ? section.lowerPosition : validator.chunkStart(section.lowerPosition);
// if the transfer does not start on the valididator's chunk boundary, this is the number of bytes to offset by
int transferOffset = (int) (section.lowerPosition - start);
if (validator != null)
validator.seek(start);
// length of the section to read
long length = section.upperPosition - start;
// tracks write progress
long bytesRead = 0;
while (bytesRead < length)
{
int toTransfer = (int) Math.min(bufferSize, length - bytesRead);
long lastBytesRead = write(proxy, validator, out, start, transferOffset, toTransfer, bufferSize);
start += lastBytesRead;
bytesRead += lastBytesRead;
long delta = lastBytesRead - transferOffset;
progress += delta;
session.progress(filename, ProgressInfo.Direction.OUT, progress, delta, totalSize);
transferOffset = 0;
}
// make sure that current section is sent
out.flush();
}
logger.debug("[Stream #{}] Finished streaming file {} to {}, bytesTransferred = {}, totalSize = {}",
session.planId(), sstable.getFilename(), session.peer, FBUtilities.prettyPrintMemory(progress), FBUtilities.prettyPrintMemory(totalSize));
}
}
protected long totalSize()
{
return totalSize;
}
/**
* Sequentially read bytes from the file and write them to the output stream
*
* @param proxy The file reader to read from
* @param validator validator to verify data integrity
* @param start The readd offset from the beginning of the {@code proxy} file.
* @param transferOffset number of bytes to skip transfer, but include for validation.
* @param toTransfer The number of bytes to be transferred.
*
* @return Number of bytes transferred.
*
* @throws java.io.IOException on any I/O error
*/
protected long write(ChannelProxy proxy, ChecksumValidator validator, StreamingDataOutputPlus output, long start, int transferOffset, int toTransfer, int bufferSize) throws IOException
{
// the count of bytes to read off disk
int minReadable = (int) Math.min(bufferSize, proxy.size() - start);
// this buffer will hold the data from disk. as it will be compressed on the fly by
// AsyncChannelCompressedStreamWriter.write(ByteBuffer), we can release this buffer as soon as we can.
ByteBuffer buffer = BufferPools.forNetworking().get(minReadable, BufferType.OFF_HEAP);
try
{
int readCount = proxy.read(buffer, start);
assert readCount == minReadable : String.format("could not read required number of bytes from file to be streamed: read %d bytes, wanted %d bytes", readCount, minReadable);
buffer.flip();
if (validator != null)
{
validator.validate(buffer);
buffer.flip();
}
buffer.position(transferOffset);
buffer.limit(transferOffset + (toTransfer - transferOffset));
output.writeToChannel(StreamCompressionSerializer.serialize(compressor, buffer, current_version), limiter);
}
finally
{
BufferPools.forNetworking().put(buffer);
}
return toTransfer;
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy