io.trino.filesystem.hdfs.HdfsOutputFile Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of trino-hdfs Show documentation
Show all versions of trino-hdfs Show documentation
Trino - Legacy HDFS file system support
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package io.trino.filesystem.hdfs;
import com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem;
import io.airlift.stats.TimeStat;
import io.trino.filesystem.Location;
import io.trino.filesystem.TrinoOutputFile;
import io.trino.hdfs.CallStats;
import io.trino.hdfs.HdfsContext;
import io.trino.hdfs.HdfsEnvironment;
import io.trino.hdfs.MemoryAwareFileSystem;
import io.trino.hdfs.authentication.HdfsAuthentication.ExceptionAction;
import io.trino.hdfs.gcs.GcsAtomicOutputStream;
import io.trino.memory.context.AggregatedMemoryContext;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import java.io.IOException;
import java.io.OutputStream;
import java.nio.file.FileAlreadyExistsException;
import static io.trino.filesystem.hdfs.HadoopPaths.hadoopPath;
import static io.trino.filesystem.hdfs.HdfsFileSystem.withCause;
import static io.trino.hdfs.FileSystemUtils.getRawFileSystem;
import static io.trino.memory.context.AggregatedMemoryContext.newSimpleAggregatedMemoryContext;
import static java.util.Objects.requireNonNull;
class HdfsOutputFile
implements TrinoOutputFile
{
private final Location location;
private final HdfsEnvironment environment;
private final HdfsContext context;
private final CallStats createFileCallStat;
public HdfsOutputFile(Location location, HdfsEnvironment environment, HdfsContext context, CallStats createFileCallStat)
{
this.location = requireNonNull(location, "location is null");
this.environment = requireNonNull(environment, "environment is null");
this.context = requireNonNull(context, "context is null");
this.createFileCallStat = requireNonNull(createFileCallStat, "createFileCallStat is null");
location.verifyValidFileLocation();
}
@Override
public OutputStream create(AggregatedMemoryContext memoryContext)
throws IOException
{
return create(false, memoryContext);
}
@Override
public void createOrOverwrite(byte[] data)
throws IOException
{
try (OutputStream out = create(true, newSimpleAggregatedMemoryContext())) {
out.write(data);
}
}
@Override
public void createExclusive(byte[] data)
throws IOException
{
Path file = hadoopPath(location);
FileSystem fileSystem = getRawFileSystem(environment.getFileSystem(context, file));
if (fileSystem instanceof GoogleHadoopFileSystem) {
GcsAtomicOutputStream atomicOutputStream = new GcsAtomicOutputStream(environment, context, file);
atomicOutputStream.write(data);
atomicOutputStream.close();
return;
}
throw new UnsupportedOperationException("createExclusive not supported for " + fileSystem);
}
private OutputStream create(boolean overwrite, AggregatedMemoryContext memoryContext)
throws IOException
{
createFileCallStat.newCall();
Path file = hadoopPath(location);
FileSystem fileSystem = environment.getFileSystem(context, file);
FileSystem rawFileSystem = getRawFileSystem(fileSystem);
try (TimeStat.BlockTimer _ = createFileCallStat.time()) {
if (rawFileSystem instanceof MemoryAwareFileSystem memoryAwareFileSystem) {
return create(() -> memoryAwareFileSystem.create(file, memoryContext));
}
return create(() -> fileSystem.create(file, overwrite));
}
catch (org.apache.hadoop.fs.FileAlreadyExistsException e) {
createFileCallStat.recordException(e);
throw withCause(new FileAlreadyExistsException(toString()), e);
}
catch (IOException e) {
createFileCallStat.recordException(e);
throw new IOException("Creation of file %s failed: %s".formatted(file, e.getMessage()), e);
}
}
private OutputStream create(ExceptionAction action)
throws IOException
{
FSDataOutputStream out = environment.doAs(context.getIdentity(), action);
return new HdfsOutputStream(location, out, environment, context);
}
@Override
public Location location()
{
return location;
}
@Override
public String toString()
{
return location().toString();
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy