org.apache.cassandra.db.streaming.CassandraOutgoingFile Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of cassandra-all Show documentation
Show all versions of cassandra-all Show documentation
The Apache Cassandra Project develops a highly scalable second-generation distributed database, bringing together Dynamo's fully distributed design and Bigtable's ColumnFamily-based data model.
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.cassandra.db.streaming;
import java.io.IOException;
import java.util.List;
import java.util.Objects;
import java.util.UUID;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Preconditions;
import org.apache.cassandra.config.DatabaseDescriptor;
import org.apache.cassandra.dht.Range;
import org.apache.cassandra.dht.Token;
import org.apache.cassandra.io.sstable.format.SSTableReader;
import org.apache.cassandra.io.util.DataOutputStreamPlus;
import org.apache.cassandra.net.AsyncStreamingOutputPlus;
import org.apache.cassandra.schema.TableId;
import org.apache.cassandra.streaming.OutgoingStream;
import org.apache.cassandra.streaming.StreamOperation;
import org.apache.cassandra.streaming.StreamSession;
import org.apache.cassandra.utils.concurrent.Ref;
/**
* used to transfer the part(or whole) of a SSTable data file
*/
public class CassandraOutgoingFile implements OutgoingStream
{
private final Ref ref;
private final long estimatedKeys;
private final List sections;
private final String filename;
private final boolean shouldStreamEntireSSTable;
private final StreamOperation operation;
private final CassandraStreamHeader header;
public CassandraOutgoingFile(StreamOperation operation, Ref ref,
List sections, List> normalizedRanges,
long estimatedKeys)
{
Preconditions.checkNotNull(ref.get());
Range.assertNormalized(normalizedRanges);
this.operation = operation;
this.ref = ref;
this.estimatedKeys = estimatedKeys;
this.sections = sections;
SSTableReader sstable = ref.get();
this.filename = sstable.getFilename();
this.shouldStreamEntireSSTable = computeShouldStreamEntireSSTables();
ComponentManifest manifest = ComponentManifest.create(sstable.descriptor);
this.header = makeHeader(sstable, operation, sections, estimatedKeys, shouldStreamEntireSSTable, manifest);
}
private static CassandraStreamHeader makeHeader(SSTableReader sstable,
StreamOperation operation,
List sections,
long estimatedKeys,
boolean shouldStreamEntireSSTable,
ComponentManifest manifest)
{
boolean keepSSTableLevel = operation == StreamOperation.BOOTSTRAP || operation == StreamOperation.REBUILD;
CompressionInfo compressionInfo = sstable.compression
? CompressionInfo.newLazyInstance(sstable.getCompressionMetadata(), sections)
: null;
return CassandraStreamHeader.builder()
.withSSTableFormat(sstable.descriptor.formatType)
.withSSTableVersion(sstable.descriptor.version)
.withSSTableLevel(keepSSTableLevel ? sstable.getSSTableLevel() : 0)
.withEstimatedKeys(estimatedKeys)
.withSections(sections)
.withCompressionInfo(compressionInfo)
.withSerializationHeader(sstable.header.toComponent())
.isEntireSSTable(shouldStreamEntireSSTable)
.withComponentManifest(manifest)
.withFirstKey(sstable.first)
.withTableId(sstable.metadata().id)
.build();
}
@VisibleForTesting
public static CassandraOutgoingFile fromStream(OutgoingStream stream)
{
Preconditions.checkArgument(stream instanceof CassandraOutgoingFile);
return (CassandraOutgoingFile) stream;
}
@VisibleForTesting
public Ref getRef()
{
return ref;
}
@Override
public String getName()
{
return filename;
}
@Override
public long getEstimatedSize()
{
return header.size();
}
@Override
public TableId getTableId()
{
return ref.get().metadata().id;
}
@Override
public int getNumFiles()
{
return shouldStreamEntireSSTable ? header.componentManifest.components().size() : 1;
}
@Override
public long getRepairedAt()
{
return ref.get().getRepairedAt();
}
@Override
public UUID getPendingRepair()
{
return ref.get().getPendingRepair();
}
@Override
public void write(StreamSession session, DataOutputStreamPlus out, int version) throws IOException
{
// FileStreamTask uses AsyncStreamingOutputPlus for streaming.
assert out instanceof AsyncStreamingOutputPlus : "Unexpected DataOutputStreamPlus " + out.getClass();
SSTableReader sstable = ref.get();
if (shouldStreamEntireSSTable)
{
// Acquire lock to avoid concurrent sstable component mutation because of stats update or index summary
// redistribution, otherwise file sizes recorded in component manifest will be different from actual
// file sizes. (Note: Windows doesn't support atomic replace and index summary redistribution deletes
// existing file first)
// Recreate the latest manifest and hard links for mutatable components in case they are modified.
try (ComponentContext context = sstable.runWithLock(ignored -> ComponentContext.create(sstable.descriptor)))
{
CassandraStreamHeader current = makeHeader(sstable, operation, sections, estimatedKeys, true, context.manifest());
CassandraStreamHeader.serializer.serialize(current, out, version);
out.flush();
CassandraEntireSSTableStreamWriter writer = new CassandraEntireSSTableStreamWriter(sstable, session, context);
writer.write((AsyncStreamingOutputPlus) out);
}
}
else
{
// legacy streaming is not affected by stats metadata mutation and index sumary redistribution
CassandraStreamHeader.serializer.serialize(header, out, version);
out.flush();
CassandraStreamWriter writer = header.isCompressed() ?
new CassandraCompressedStreamWriter(sstable, header, session) :
new CassandraStreamWriter(sstable, header, session);
writer.write(out);
}
}
@VisibleForTesting
public boolean computeShouldStreamEntireSSTables()
{
// don't stream if full sstable transfers are disabled or legacy counter shards are present
if (!DatabaseDescriptor.streamEntireSSTables() || ref.get().getSSTableMetadata().hasLegacyCounterShards)
return false;
return contained(sections, ref.get());
}
@VisibleForTesting
public boolean contained(List sections, SSTableReader sstable)
{
if (sections == null || sections.isEmpty())
return false;
// if transfer sections contain entire sstable
long transferLength = sections.stream().mapToLong(p -> p.upperPosition - p.lowerPosition).sum();
return transferLength == sstable.uncompressedLength();
}
@Override
public void finish()
{
ref.release();
}
public boolean equals(Object o)
{
if (this == o) return true;
if (o == null || getClass() != o.getClass()) return false;
CassandraOutgoingFile that = (CassandraOutgoingFile) o;
return estimatedKeys == that.estimatedKeys &&
Objects.equals(ref, that.ref) &&
Objects.equals(sections, that.sections);
}
public int hashCode()
{
return Objects.hash(ref, estimatedKeys, sections);
}
@Override
public String toString()
{
return "CassandraOutgoingFile{" + filename + '}';
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy