All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.datatorrent.lib.io.fs.FSInputModule Maven / Gradle / Ivy

/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *   http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */
package com.datatorrent.lib.io.fs;

import javax.validation.constraints.Min;
import javax.validation.constraints.NotNull;
import javax.validation.constraints.Size;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context;
import com.datatorrent.api.DAG;
import com.datatorrent.api.Module;
import com.datatorrent.common.partitioner.StatelessPartitioner;
import com.datatorrent.lib.codec.KryoSerializableStreamCodec;
import com.datatorrent.lib.io.block.AbstractBlockReader;
import com.datatorrent.lib.io.block.BlockMetadata;
import com.datatorrent.lib.io.block.FSSliceReader;
import com.datatorrent.netlet.util.Slice;

/**
 * FSInputModule is an abstract class used to read files from file systems like HDFS, NFS, S3, etc. 
* FSInputModule emits FileMetadata, BlockMetadata, BlockBytes.
* The module reads data in parallel, following parameters can be configured
* 1. files: list of file(s)/directories to read
* 2. filePatternRegularExp: Files names matching given regex will be read
* 3. scanIntervalMillis: interval between two scans to discover new files in input directory
* 4. recursive: if scan recursively input directories
* 5. blockSize: block size used to read input blocks of file
* 6. readersCount: count of readers to read input file
* 7. sequencialFileRead: If emit file blocks in sequence? * * @since 3.4.0 */ public class FSInputModule implements Module { @NotNull @Size(min = 1) private String files; private String filePatternRegularExp; @Min(0) private long scanIntervalMillis; private boolean recursive = true; private long blockSize; private boolean sequencialFileRead = false; private int readersCount; public final transient ProxyOutputPort filesMetadataOutput = new ProxyOutputPort<>(); public final transient ProxyOutputPort blocksMetadataOutput = new ProxyOutputPort<>(); public final transient ProxyOutputPort> messages = new ProxyOutputPort<>(); public FileSplitterInput createFileSplitter() { return new FileSplitterInput(); } public FSSliceReader createBlockReader() { return new FSSliceReader(); } @Override public void populateDAG(DAG dag, Configuration conf) { FileSplitterInput fileSplitter = dag.addOperator("FileSplitter", createFileSplitter()); FSSliceReader blockReader = dag.addOperator("BlockReader", createBlockReader()); dag.addStream("BlockMetadata", fileSplitter.blocksMetadataOutput, blockReader.blocksMetadataInput); filesMetadataOutput.set(fileSplitter.filesMetadataOutput); blocksMetadataOutput.set(blockReader.blocksMetadataOutput); messages.set(blockReader.messages); if (sequencialFileRead) { dag.setInputPortAttribute(blockReader.blocksMetadataInput, Context.PortContext.STREAM_CODEC, new SequentialFileBlockMetadataCodec()); } if (blockSize != 0) { fileSplitter.setBlockSize(blockSize); } FileSplitterInput.TimeBasedDirectoryScanner fileScanner = fileSplitter.getScanner(); fileScanner.setFiles(files); if (scanIntervalMillis != 0) { fileScanner.setScanIntervalMillis(scanIntervalMillis); } fileScanner.setRecursive(recursive); if (filePatternRegularExp != null) { fileSplitter.getScanner().setFilePatternRegularExp(filePatternRegularExp); } blockReader.setBasePath(files); if (readersCount != 0) { dag.setAttribute(blockReader, Context.OperatorContext.PARTITIONER, new StatelessPartitioner(readersCount)); fileSplitter.setBlocksThreshold(readersCount); } } /** * A comma separated list of directories to scan. If the path is not fully qualified the default file system is used. * A fully qualified path can be provided to scan directories in other filesystems. * * @param files * files */ public void setFiles(String files) { this.files = files; } /** * Gets the files to be scanned. * * @return files to be scanned. */ public String getFiles() { return files; } /** * Gets the regular expression for file names to split * * @return regular expression */ public String getFilePatternRegularExp() { return filePatternRegularExp; } /** * Only files with names matching the given java regular expression are split * * @param filePatternRegexp * regular expression */ public void setFilePatternRegularExp(String filePatternRegexp) { this.filePatternRegularExp = filePatternRegexp; } /** * Gets scan interval in milliseconds, interval between two scans to discover new files in input directory * * @return scanInterval milliseconds */ public long getScanIntervalMillis() { return scanIntervalMillis; } /** * Sets scan interval in milliseconds, interval between two scans to discover new files in input directory * * @param scanIntervalMillis */ public void setScanIntervalMillis(long scanIntervalMillis) { this.scanIntervalMillis = scanIntervalMillis; } /** * Get is scan recursive * * @return isRecursive */ public boolean isRecursive() { return recursive; } /** * set is scan recursive * * @param recursive */ public void setRecursive(boolean recursive) { this.recursive = recursive; } /** * Get block size used to read input blocks of file * * @return blockSize */ public long getBlockSize() { return blockSize; } /** * Sets block size used to read input blocks of file * * @param blockSize */ public void setBlockSize(long blockSize) { this.blockSize = blockSize; } /** * Gets readers count * @return readersCount */ public int getReadersCount() { return readersCount; } /** * Static count of readers to read input file * @param readersCount */ public void setReadersCount(int readersCount) { this.readersCount = readersCount; } /** * Gets is sequencial file read * * @return sequencialFileRead */ public boolean isSequencialFileRead() { return sequencialFileRead; } /** * Sets is sequencial file read * * @param sequencialFileRead */ public void setSequencialFileRead(boolean sequencialFileRead) { this.sequencialFileRead = sequencialFileRead; } public static class SequentialFileBlockMetadataCodec extends KryoSerializableStreamCodec { @Override public int getPartition(BlockMetadata.FileBlockMetadata fileBlockMetadata) { return fileBlockMetadata.hashCode(); } } }




© 2015 - 2025 Weber Informatics LLC | Privacy Policy