All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.swirlds.merkledb.files.DataFileIterator Maven / Gradle / Ivy

Go to download

Swirlds is a software platform designed to build fully-distributed applications that harness the power of the cloud without servers. Now you can develop applications with fairness in decision making, speed, trust and reliability, at a fraction of the cost of traditional server-based platforms.

There is a newer version: 0.56.6
Show newest version
/*
 * Copyright (C) 2023-2024 Hedera Hashgraph, LLC
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *      http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package com.swirlds.merkledb.files;

import static com.hedera.pbj.runtime.ProtoParserTools.TAG_FIELD_OFFSET;
import static com.swirlds.merkledb.files.DataFileCommon.FIELD_DATAFILE_ITEMS;
import static com.swirlds.merkledb.files.DataFileCommon.FIELD_DATAFILE_METADATA;

import com.hedera.pbj.runtime.io.ReadableSequentialData;
import com.hedera.pbj.runtime.io.buffer.BufferedData;
import com.hedera.pbj.runtime.io.stream.ReadableStreamingData;
import com.swirlds.base.utility.ToStringBuilder;
import com.swirlds.merkledb.config.MerkleDbConfig;
import java.io.BufferedInputStream;
import java.io.IOException;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.StandardOpenOption;
import java.util.Objects;

/**
 * Iterator class for iterating over data items in a DataFile. It is designed to be used in a while(iter.next()){...}
 * loop and you can then read the data items info for current item with getDataItemsKey, getDataItemsDataLocation and
 * getDataItemData.
 *
 * 

It is designed to be used from a single thread. * * @see DataFileReader for definition of file structure */ public final class DataFileIterator implements AutoCloseable { /** Input stream this iterator is reading from */ private final BufferedInputStream inputStream; /** Readable data on top of the input stream */ private final ReadableSequentialData in; /** The file metadata read from the end of file */ private final DataFileMetadata metadata; /** The path to the file we are iterating over */ private final Path path; /** Buffer that is reused for reading each data item */ private BufferedData dataItemBuffer; /** Index of current data item this iterator is reading, zero being the first item, -1 being before start */ private long currentDataItem = -1; /** The offset in bytes from start of file to the beginning of the current item. */ private long currentDataItemFilePosition = 0; /** True if this iterator has been closed */ private boolean closed = false; /** * Create a new DataFileIterator on an existing file. * * @param dbConfig MerkleDb config * @param path * The path to the file to read. * @param metadata * The metadata read from the file. * @throws IOException * if there was a problem creating a new InputStream on the file at path */ public DataFileIterator(final MerkleDbConfig dbConfig, final Path path, final DataFileMetadata metadata) throws IOException { this.path = path; this.metadata = metadata; this.inputStream = new BufferedInputStream( Files.newInputStream(path, StandardOpenOption.READ), dbConfig.iteratorInputBufferBytes()); this.in = new ReadableStreamingData(inputStream); this.in.limit(Files.size(path)); } /** * Get the path for the data file. */ public Path getPath() { return path; } /** * Get the metadata for the data file. * * @return File's metadata */ public DataFileMetadata getMetadata() { return metadata; } /** * Close the iterator. * * @throws IOException if this resource cannot be closed */ @Override public void close() throws IOException { if (!closed) { closed = true; dataItemBuffer = null; inputStream.close(); } } /** * Advance to the next dataItem. * * @return true if a dataItem was read or false if the end of the file has been reached. * @throws IOException * If there was a problem reading from file. */ public boolean next() throws IOException { if (closed) { throw new IllegalStateException("Cannot read from a closed iterator"); } // Have we reached the end? if (currentDataItem >= metadata.getDataItemCount() - 1) { dataItemBuffer = null; return false; } while (in.hasRemaining()) { currentDataItemFilePosition = in.position(); final int tag = in.readVarInt(false); final int fieldNum = tag >> TAG_FIELD_OFFSET; if (fieldNum == FIELD_DATAFILE_ITEMS.number()) { final int currentDataItemSize = in.readVarInt(false); dataItemBuffer = fillBuffer(currentDataItemSize); currentDataItem++; return true; } else if (fieldNum == FIELD_DATAFILE_METADATA.number()) { final int metadataSize = in.readVarInt(false); in.skip(metadataSize); } else { throw new IllegalArgumentException("Unknown data file field: " + fieldNum); } } throw new IllegalStateException("Reached the end of data file while expecting more data items"); } /** * Get the current dataItems data. This is a shared buffer and must NOT be leaked from * the call site or modified directly. * * @return buffer containing the key and value data. This will return null if the iterator has * been closed, or if the iterator is in the before-first or after-last states. */ public BufferedData getDataItemData() { return dataItemBuffer; } /** * Get the data location (file + offset) for the current data item. * * @return current data item location */ public long getDataItemDataLocation() { return DataFileCommon.dataLocation(metadata.getIndex(), currentDataItemFilePosition); } /** toString for debugging */ @Override public String toString() { return new ToStringBuilder(this) .append("fileIndex", metadata.getIndex()) .append("currentDataItemIndex", currentDataItem) .append("currentDataItemByteOffset", currentDataItemFilePosition) .append("fileName", path.getFileName()) .append("metadata", metadata) .toString(); } /** * Equals for use when comparing in collections, based on matching file paths and metadata */ @Override public boolean equals(final Object o) { if (this == o) { return true; } if (o == null || getClass() != o.getClass()) { return false; } final DataFileIterator that = (DataFileIterator) o; return path.equals(that.getPath()) && metadata.equals(that.getMetadata()); } /** * hashCode for use when comparing in collections, based on file path and metadata */ @Override public int hashCode() { return Objects.hash(path, metadata); } // ================================================================================================================= // Private methods /** * Reads bytesToRead bytes from the current data item * @param bytesToRead bytes to read * @return ByteBuffer containing requested bytes * @throws IOException if request can not be completed */ private BufferedData fillBuffer(int bytesToRead) throws IOException { if (bytesToRead <= 0) { throw new IOException("Malformed file [" + path + "], data item [" + currentDataItem + "], requested bytes [" + bytesToRead + "]"); } // Create or resize the buffer if necessary if (dataItemBuffer == null || dataItemBuffer.capacity() < bytesToRead) { resizeBuffer(bytesToRead); } dataItemBuffer.position(0); dataItemBuffer.limit(bytesToRead); final long bytesRead = in.readBytes(dataItemBuffer); if (bytesRead != bytesToRead) { throw new IOException("Couldn't read " + bytesToRead + " bytes, data item [" + currentDataItem + "], requested bytes [" + bytesToRead + "]"); } dataItemBuffer.position(0); return dataItemBuffer; } /** * Resizes the dataItemBuffer, or creates it if necessary, such that it is large enough * to read the bytes provided. * * @param bytesToRead * Number of bytes to be able to fit into the buffer. */ private void resizeBuffer(int bytesToRead) { dataItemBuffer = BufferedData.allocate(bytesToRead); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy