io.trino.plugin.exchange.hdfs.HadoopFileSystemExchangeStorage Maven / Gradle / Ivy
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package io.trino.plugin.exchange.hdfs;
import com.google.common.collect.ImmutableList;
import com.google.common.util.concurrent.ListenableFuture;
import com.google.errorprone.annotations.ThreadSafe;
import com.google.errorprone.annotations.concurrent.GuardedBy;
import com.google.inject.Inject;
import io.airlift.slice.InputStreamSliceInput;
import io.airlift.slice.Slice;
import io.trino.annotation.NotThreadSafe;
import io.trino.plugin.exchange.filesystem.ExchangeSourceFile;
import io.trino.plugin.exchange.filesystem.ExchangeStorageReader;
import io.trino.plugin.exchange.filesystem.ExchangeStorageWriter;
import io.trino.plugin.exchange.filesystem.FileStatus;
import io.trino.plugin.exchange.filesystem.FileSystemExchangeStorage;
import io.trino.spi.classloader.ThreadContextClassLoader;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.LocatedFileStatus;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.RemoteIterator;
import java.io.File;
import java.io.IOException;
import java.io.OutputStream;
import java.io.UncheckedIOException;
import java.net.URI;
import java.util.ArrayDeque;
import java.util.List;
import java.util.Queue;
import static com.google.common.base.Preconditions.checkArgument;
import static com.google.common.util.concurrent.Futures.immediateFailedFuture;
import static com.google.common.util.concurrent.Futures.immediateFuture;
import static com.google.common.util.concurrent.Futures.immediateVoidFuture;
import static io.airlift.slice.SizeOf.instanceSize;
import static java.lang.Math.toIntExact;
import static java.util.Objects.requireNonNull;
public class HadoopFileSystemExchangeStorage
implements FileSystemExchangeStorage
{
private final int blockSize;
private final FileSystem fileSystem;
@Inject
public HadoopFileSystemExchangeStorage(ExchangeHdfsConfig config)
throws IOException
{
Configuration hdfsConfig = newEmptyConfiguration();
for (File resourcePath : config.getResourceConfigFiles()) {
checkArgument(resourcePath.exists(), "File does not exist: %s", resourcePath);
hdfsConfig.addResource(new Path(resourcePath.getPath()));
}
fileSystem = FileSystem.get(hdfsConfig);
blockSize = toIntExact(config.getHdfsStorageBlockSize().toBytes());
}
@Override
public void createDirectories(URI dir)
throws IOException
{
fileSystem.mkdirs(new Path(dir));
}
@Override
public ExchangeStorageReader createExchangeStorageReader(List sourceFiles, int maxPageStorageSize)
{
return new HadoopExchangeStorageReader(fileSystem, sourceFiles, blockSize);
}
@Override
public ExchangeStorageWriter createExchangeStorageWriter(URI file)
{
return new HadoopExchangeStorageWriter(fileSystem, file);
}
@Override
public ListenableFuture createEmptyFile(URI file)
{
try {
fileSystem.createNewFile(new Path(file));
}
catch (IOException e) {
return immediateFailedFuture(e);
}
return immediateVoidFuture();
}
@Override
public ListenableFuture deleteRecursively(List directories)
{
for (URI dir : directories) {
try {
fileSystem.delete(new Path(dir), true);
}
catch (IOException | RuntimeException e) {
return immediateFailedFuture(e);
}
}
return immediateVoidFuture();
}
@Override
public ListenableFuture> listFilesRecursively(URI dir)
{
ImmutableList.Builder builder = ImmutableList.builder();
try {
RemoteIterator fileStatusListIterator = fileSystem.listFiles(
new Path(dir), true);
while (fileStatusListIterator.hasNext()) {
LocatedFileStatus fileStatus = fileStatusListIterator.next();
builder.add(new FileStatus(fileStatus.getPath().toString(), fileStatus.getLen()));
}
}
catch (IOException e) {
return immediateFailedFuture(e);
}
return immediateFuture(builder.build());
}
@Override
public int getWriteBufferSize()
{
return blockSize;
}
@Override
public void close()
{
}
private static Configuration newEmptyConfiguration()
{
try (ThreadContextClassLoader ignored = new ThreadContextClassLoader(HadoopFileSystemExchangeStorage.class.getClassLoader())) {
return new Configuration(false);
}
}
@ThreadSafe
private static class HadoopExchangeStorageReader
implements ExchangeStorageReader
{
private static final int INSTANCE_SIZE = instanceSize(HadoopExchangeStorageReader.class);
private final FileSystem fileSystem;
@GuardedBy("this")
private final Queue sourceFiles;
private final int blockSize;
@GuardedBy("this")
private InputStreamSliceInput sliceInput;
@GuardedBy("this")
private boolean closed;
public HadoopExchangeStorageReader(FileSystem fileSystem, List sourceFiles, int blockSize)
{
this.fileSystem = requireNonNull(fileSystem, "fileSystem is null");
this.sourceFiles = new ArrayDeque<>(requireNonNull(sourceFiles, "sourceFiles is null"));
this.blockSize = blockSize;
}
@Override
public synchronized Slice read()
throws IOException
{
if (closed) {
return null;
}
if (sliceInput != null) {
if (sliceInput.isReadable()) {
return sliceInput.readSlice(sliceInput.readInt());
}
else {
sliceInput.close();
}
}
ExchangeSourceFile sourceFile = sourceFiles.poll();
if (sourceFile == null) {
close();
return null;
}
sliceInput = getSliceInput(sourceFile);
return sliceInput.readSlice(sliceInput.readInt());
}
@Override
public ListenableFuture isBlocked()
{
return immediateVoidFuture();
}
@Override
public synchronized long getRetainedSize()
{
return INSTANCE_SIZE + (sliceInput == null ? 0 : sliceInput.getRetainedSize());
}
@Override
public synchronized boolean isFinished()
{
return closed;
}
@Override
public synchronized void close()
{
if (closed) {
return;
}
closed = true;
if (sliceInput != null) {
sliceInput.close();
sliceInput = null;
}
}
private InputStreamSliceInput getSliceInput(ExchangeSourceFile sourceFile)
throws IOException
{
Path fileURL = new Path(sourceFile.getFileUri());
return new InputStreamSliceInput(fileSystem.open(fileURL), blockSize);
}
}
@NotThreadSafe
private static class HadoopExchangeStorageWriter
implements ExchangeStorageWriter
{
private static final int INSTANCE_SIZE = instanceSize(HadoopExchangeStorageReader.class);
private final OutputStream outputStream;
public HadoopExchangeStorageWriter(FileSystem fileSystem, URI file)
{
try {
this.outputStream = fileSystem.create(new Path(file), true);
}
catch (IOException e) {
throw new UncheckedIOException(e);
}
}
@Override
public ListenableFuture write(Slice slice)
{
try {
outputStream.write(slice.getBytes());
}
catch (IOException | RuntimeException e) {
return immediateFailedFuture(e);
}
return immediateVoidFuture();
}
@Override
public ListenableFuture finish()
{
try {
outputStream.close();
}
catch (IOException | RuntimeException e) {
return immediateFailedFuture(e);
}
return immediateVoidFuture();
}
@Override
public ListenableFuture abort()
{
try {
outputStream.close();
}
catch (IOException | RuntimeException e) {
return immediateFailedFuture(e);
}
return immediateVoidFuture();
}
@Override
public long getRetainedSize()
{
return INSTANCE_SIZE;
}
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy