All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.hbase.util.FSTableDescriptors Maven / Gradle / Ivy

There is a newer version: 3.0.0-beta-1
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hbase.util;

import com.google.errorprone.annotations.RestrictedApi;
import edu.umd.cs.findbugs.annotations.Nullable;
import java.io.EOFException;
import java.io.IOException;
import java.io.InterruptedIOException;
import java.util.Arrays;
import java.util.Comparator;
import java.util.List;
import java.util.Map;
import java.util.Optional;
import java.util.TreeMap;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.ConcurrentSkipListMap;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.LinkedBlockingQueue;
import java.util.concurrent.ThreadPoolExecutor;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicBoolean;
import org.apache.commons.lang3.NotImplementedException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileAlreadyExistsException;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter;
import org.apache.hadoop.hbase.Coprocessor;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.TableDescriptors;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
import org.apache.hadoop.hbase.client.CoprocessorDescriptorBuilder;
import org.apache.hadoop.hbase.client.TableDescriptor;
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
import org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint;
import org.apache.hadoop.hbase.exceptions.DeserializationException;
import org.apache.hadoop.hbase.regionserver.BloomType;
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
import org.apache.yetus.audience.InterfaceAudience;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import org.apache.hbase.thirdparty.com.google.common.primitives.Ints;
import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;

/**
 * Implementation of {@link TableDescriptors} that reads descriptors from the passed filesystem. It
 * expects descriptors to be in a file in the {@link #TABLEINFO_DIR} subdir of the table's directory
 * in FS. Can be read-only -- i.e. does not modify the filesystem or can be read and write.
 * 

* Also has utility for keeping up the table descriptors tableinfo file. The table schema file is * kept in the {@link #TABLEINFO_DIR} subdir of the table directory in the filesystem. It has a * {@link #TABLEINFO_FILE_PREFIX} and then a suffix that is the edit sequenceid: e.g. * .tableinfo.0000000003. This sequenceid is always increasing. It starts at zero. The * table schema file with the highest sequenceid has the most recent schema edit. Usually there is * one file only, the most recent but there may be short periods where there are more than one file. * Old files are eventually cleaned. Presumption is that there will not be lots of concurrent * clients making table schema edits. If so, the below needs a bit of a reworking and perhaps some * supporting api in hdfs. */ @InterfaceAudience.Private public class FSTableDescriptors implements TableDescriptors { private static final Logger LOG = LoggerFactory.getLogger(FSTableDescriptors.class); private final FileSystem fs; private final Path rootdir; private final boolean fsreadonly; private final boolean usecache; private volatile boolean fsvisited; private boolean tableDescriptorParallelLoadEnable = false; private ThreadPoolExecutor executor; long cachehits = 0; long invocations = 0; /** * The file name prefix used to store HTD in HDFS */ static final String TABLEINFO_FILE_PREFIX = ".tableinfo"; public static final String TABLEINFO_DIR = ".tabledesc"; // This cache does not age out the old stuff. Thinking is that the amount // of data we keep up in here is so small, no need to do occasional purge. // TODO. private final Map cache = new ConcurrentHashMap<>(); /** * Construct a FSTableDescriptors instance using the hbase root dir of the given conf and the * filesystem where that root dir lives. This instance can do write operations (is not read only). */ public FSTableDescriptors(final Configuration conf) throws IOException { this(CommonFSUtils.getCurrentFileSystem(conf), CommonFSUtils.getRootDir(conf)); } public FSTableDescriptors(final FileSystem fs, final Path rootdir) { this(fs, rootdir, false, true); } public FSTableDescriptors(final FileSystem fs, final Path rootdir, final boolean fsreadonly, final boolean usecache) { this(fs, rootdir, fsreadonly, usecache, 0); } public FSTableDescriptors(final FileSystem fs, final Path rootdir, final boolean fsreadonly, final boolean usecache, final int tableDescriptorParallelLoadThreads) { this.fs = fs; this.rootdir = rootdir; this.fsreadonly = fsreadonly; this.usecache = usecache; if (tableDescriptorParallelLoadThreads > 0) { tableDescriptorParallelLoadEnable = true; executor = new ThreadPoolExecutor(tableDescriptorParallelLoadThreads, tableDescriptorParallelLoadThreads, 1, TimeUnit.SECONDS, new LinkedBlockingQueue<>(), new ThreadFactoryBuilder().setNameFormat("FSTableDescriptorLoad-pool-%d").setDaemon(true) .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); executor.allowCoreThreadTimeOut(true); } } public static void tryUpdateMetaTableDescriptor(Configuration conf) throws IOException { tryUpdateAndGetMetaTableDescriptor(conf, CommonFSUtils.getCurrentFileSystem(conf), CommonFSUtils.getRootDir(conf)); } public static TableDescriptor tryUpdateAndGetMetaTableDescriptor(Configuration conf, FileSystem fs, Path rootdir) throws IOException { // see if we already have meta descriptor on fs. Write one if not. Optional> opt = getTableDescriptorFromFs(fs, CommonFSUtils.getTableDir(rootdir, TableName.META_TABLE_NAME), false); if (opt.isPresent()) { return opt.get().getSecond(); } TableDescriptorBuilder builder = createMetaTableDescriptorBuilder(conf); TableDescriptor td = StoreFileTrackerFactory.updateWithTrackerConfigs(conf, builder.build()); LOG.info("Creating new hbase:meta table descriptor {}", td); TableName tableName = td.getTableName(); Path tableDir = CommonFSUtils.getTableDir(rootdir, tableName); Path p = writeTableDescriptor(fs, td, tableDir, null); if (p == null) { throw new IOException("Failed update hbase:meta table descriptor"); } LOG.info("Updated hbase:meta table descriptor to {}", p); return td; } public static ColumnFamilyDescriptor getTableFamilyDescForMeta(final Configuration conf) { return ColumnFamilyDescriptorBuilder.newBuilder(HConstants.TABLE_FAMILY) .setMaxVersions( conf.getInt(HConstants.HBASE_META_VERSIONS, HConstants.DEFAULT_HBASE_META_VERSIONS)) .setInMemory(true).setBlocksize(8 * 1024).setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .setBloomFilterType(BloomType.ROWCOL).build(); } public static ColumnFamilyDescriptor getReplBarrierFamilyDescForMeta() { return ColumnFamilyDescriptorBuilder.newBuilder(HConstants.REPLICATION_BARRIER_FAMILY) .setMaxVersions(HConstants.ALL_VERSIONS).setInMemory(true) .setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .setBloomFilterType(BloomType.ROWCOL).build(); } private static TableDescriptorBuilder createMetaTableDescriptorBuilder(final Configuration conf) throws IOException { // TODO We used to set CacheDataInL1 for META table. When we have BucketCache in file mode, now // the META table data goes to File mode BC only. Test how that affect the system. If too much, // we have to rethink about adding back the setCacheDataInL1 for META table CFs. return TableDescriptorBuilder.newBuilder(TableName.META_TABLE_NAME) .setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(HConstants.CATALOG_FAMILY) .setMaxVersions( conf.getInt(HConstants.HBASE_META_VERSIONS, HConstants.DEFAULT_HBASE_META_VERSIONS)) .setInMemory(true) .setBlocksize( conf.getInt(HConstants.HBASE_META_BLOCK_SIZE, HConstants.DEFAULT_HBASE_META_BLOCK_SIZE)) .setScope(HConstants.REPLICATION_SCOPE_LOCAL).setBloomFilterType(BloomType.ROWCOL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .build()) .setColumnFamily(getTableFamilyDescForMeta(conf)) .setColumnFamily(getReplBarrierFamilyDescForMeta()) .setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(HConstants.NAMESPACE_FAMILY) .setMaxVersions( conf.getInt(HConstants.HBASE_META_VERSIONS, HConstants.DEFAULT_HBASE_META_VERSIONS)) .setInMemory(true) .setBlocksize( conf.getInt(HConstants.HBASE_META_BLOCK_SIZE, HConstants.DEFAULT_HBASE_META_BLOCK_SIZE)) .setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .setBloomFilterType(BloomType.ROWCOL).build()) .setCoprocessor( CoprocessorDescriptorBuilder.newBuilder(MultiRowMutationEndpoint.class.getName()) .setPriority(Coprocessor.PRIORITY_SYSTEM).build()); } protected boolean isUsecache() { return this.usecache; } /** * Get the current table descriptor for the given table, or null if none exists. *

* Uses a local cache of the descriptor but still checks the filesystem on each call if * {@link #fsvisited} is not {@code true}, i.e, we haven't done a full scan yet, to see if a newer * file has been created since the cached one was read. */ @Override @Nullable public TableDescriptor get(TableName tableName) { invocations++; if (usecache) { // Look in cache of descriptors. TableDescriptor cachedtdm = this.cache.get(tableName); if (cachedtdm != null) { cachehits++; return cachedtdm; } // we do not need to go to fs any more if (fsvisited) { return null; } } TableDescriptor tdmt = null; try { tdmt = getTableDescriptorFromFs(fs, getTableDir(tableName), fsreadonly).map(Pair::getSecond) .orElse(null); } catch (IOException ioe) { LOG.debug("Exception during readTableDecriptor. Current table name = " + tableName, ioe); } // last HTD written wins if (usecache && tdmt != null) { this.cache.put(tableName, tdmt); } return tdmt; } /** * Returns a map from table name to table descriptor for all tables. */ @Override public Map getAll() throws IOException { Map tds = new ConcurrentSkipListMap<>(); if (fsvisited) { for (Map.Entry entry : this.cache.entrySet()) { tds.put(entry.getKey().getNameWithNamespaceInclAsString(), entry.getValue()); } } else { LOG.info("Fetching table descriptors from the filesystem."); final long startTime = EnvironmentEdgeManager.currentTime(); AtomicBoolean allvisited = new AtomicBoolean(usecache); List tableDirs = FSUtils.getTableDirs(fs, rootdir); if (!tableDescriptorParallelLoadEnable) { for (Path dir : tableDirs) { internalGet(dir, tds, allvisited); } } else { CountDownLatch latch = new CountDownLatch(tableDirs.size()); for (Path dir : tableDirs) { executor.submit(new Runnable() { @Override public void run() { try { internalGet(dir, tds, allvisited); } finally { latch.countDown(); } } }); } try { latch.await(); } catch (InterruptedException ie) { throw (InterruptedIOException) new InterruptedIOException().initCause(ie); } } fsvisited = allvisited.get(); LOG.info("Fetched table descriptors(size=" + tds.size() + ") cost " + (EnvironmentEdgeManager.currentTime() - startTime) + "ms."); } return tds; } private void internalGet(Path dir, Map tds, AtomicBoolean allvisited) { TableDescriptor htd = get(CommonFSUtils.getTableName(dir)); if (htd == null) { allvisited.set(false); } else { tds.put(htd.getTableName().getNameWithNamespaceInclAsString(), htd); } } /** * Find descriptors by namespace. * @see #get(org.apache.hadoop.hbase.TableName) */ @Override public Map getByNamespace(String name) throws IOException { Map htds = new TreeMap<>(); List tableDirs = FSUtils.getLocalTableDirs(fs, CommonFSUtils.getNamespaceDir(rootdir, name)); for (Path d : tableDirs) { TableDescriptor htd = get(CommonFSUtils.getTableName(d)); if (htd == null) { continue; } htds.put(CommonFSUtils.getTableName(d).getNameAsString(), htd); } return htds; } @Override public void update(TableDescriptor td, boolean cacheOnly) throws IOException { // TODO: in fact this method will only be called at master side, so fsreadonly and usecache will // always be true. In general, we'd better have a ReadOnlyFSTableDesciptors for HRegionServer // but now, HMaster extends HRegionServer, so unless making use of generic, we can not have // different implementations for HMaster and HRegionServer. Revisit this when we make HMaster // not extend HRegionServer in the future. if (fsreadonly) { throw new UnsupportedOperationException("Cannot add a table descriptor - in read only mode"); } if (!cacheOnly) { updateTableDescriptor(td); } if (usecache) { this.cache.put(td.getTableName(), td); } } @RestrictedApi(explanation = "Should only be called in tests or self", link = "", allowedOnPath = ".*/src/test/.*|.*/FSTableDescriptors\\.java") Path updateTableDescriptor(TableDescriptor td) throws IOException { TableName tableName = td.getTableName(); Path tableDir = getTableDir(tableName); Path p = writeTableDescriptor(fs, td, tableDir, getTableDescriptorFromFs(fs, tableDir, fsreadonly).map(Pair::getFirst).orElse(null)); if (p == null) { throw new IOException("Failed update"); } LOG.info("Updated tableinfo=" + p); return p; } /** * Removes the table descriptor from the local cache and returns it. If not in read only mode, it * also deletes the entire table directory(!) from the FileSystem. */ @Override public TableDescriptor remove(final TableName tablename) throws IOException { if (fsreadonly) { throw new NotImplementedException("Cannot remove a table descriptor - in read only mode"); } Path tabledir = getTableDir(tablename); if (this.fs.exists(tabledir)) { if (!this.fs.delete(tabledir, true)) { throw new IOException("Failed delete of " + tabledir.toString()); } } TableDescriptor descriptor = this.cache.remove(tablename); return descriptor; } /** * Check whether we have a valid TableDescriptor. */ public static boolean isTableDir(FileSystem fs, Path tableDir) throws IOException { return getTableDescriptorFromFs(fs, tableDir, true).isPresent(); } /** * Compare {@link FileStatus} instances by {@link Path#getName()}. Returns in reverse order. */ static final Comparator TABLEINFO_FILESTATUS_COMPARATOR = new Comparator() { @Override public int compare(FileStatus left, FileStatus right) { return right.getPath().getName().compareTo(left.getPath().getName()); } }; /** * Return the table directory in HDFS */ private Path getTableDir(TableName tableName) { return CommonFSUtils.getTableDir(rootdir, tableName); } private static final PathFilter TABLEINFO_PATHFILTER = new PathFilter() { @Override public boolean accept(Path p) { // Accept any file that starts with TABLEINFO_NAME return p.getName().startsWith(TABLEINFO_FILE_PREFIX); } }; /** * Width of the sequenceid that is a suffix on a tableinfo file. */ static final int WIDTH_OF_SEQUENCE_ID = 10; /** * @param number Number to use as suffix. * @return Returns zero-prefixed decimal version of passed number (Does absolute in case number is * negative). */ private static String formatTableInfoSequenceId(final int number) { byte[] b = new byte[WIDTH_OF_SEQUENCE_ID]; int d = Math.abs(number); for (int i = b.length - 1; i >= 0; i--) { b[i] = (byte) ((d % 10) + '0'); d /= 10; } return Bytes.toString(b); } @Override public void close() throws IOException { // Close the executor when parallel loading enabled. if (tableDescriptorParallelLoadEnable) { this.executor.shutdown(); } } static final class SequenceIdAndFileLength { final int sequenceId; final int fileLength; SequenceIdAndFileLength(int sequenceId, int fileLength) { this.sequenceId = sequenceId; this.fileLength = fileLength; } } /** * Returns the current sequence id and file length or 0 if none found. * @param p Path to a .tableinfo file. */ @RestrictedApi(explanation = "Should only be called in tests or self", link = "", allowedOnPath = ".*/src/test/.*|.*/FSTableDescriptors\\.java") static SequenceIdAndFileLength getTableInfoSequenceIdAndFileLength(Path p) { String name = p.getName(); if (!name.startsWith(TABLEINFO_FILE_PREFIX)) { throw new IllegalArgumentException("Invalid table descriptor file name: " + name); } int firstDot = name.indexOf('.', TABLEINFO_FILE_PREFIX.length()); if (firstDot < 0) { // oldest style where we do not have both sequence id and file length return new SequenceIdAndFileLength(0, 0); } int secondDot = name.indexOf('.', firstDot + 1); if (secondDot < 0) { // old stype where we do not have file length int sequenceId = Integer.parseInt(name.substring(firstDot + 1)); return new SequenceIdAndFileLength(sequenceId, 0); } int sequenceId = Integer.parseInt(name.substring(firstDot + 1, secondDot)); int fileLength = Integer.parseInt(name.substring(secondDot + 1)); return new SequenceIdAndFileLength(sequenceId, fileLength); } /** * Returns Name of tableinfo file. */ @RestrictedApi(explanation = "Should only be called in tests or self", link = "", allowedOnPath = ".*/src/test/.*|.*/FSTableDescriptors\\.java") static String getTableInfoFileName(int sequenceId, byte[] content) { return TABLEINFO_FILE_PREFIX + "." + formatTableInfoSequenceId(sequenceId) + "." + content.length; } /** * Returns the latest table descriptor for the given table directly from the file system if it * exists, bypassing the local cache. Returns null if it's not found. */ public static TableDescriptor getTableDescriptorFromFs(FileSystem fs, Path hbaseRootDir, TableName tableName) throws IOException { Path tableDir = CommonFSUtils.getTableDir(hbaseRootDir, tableName); return getTableDescriptorFromFs(fs, tableDir); } /** * Returns the latest table descriptor for the table located at the given directory directly from * the file system if it exists. */ public static TableDescriptor getTableDescriptorFromFs(FileSystem fs, Path tableDir) throws IOException { return getTableDescriptorFromFs(fs, tableDir, true).map(Pair::getSecond).orElse(null); } private static void deleteMalformedFile(FileSystem fs, Path file) throws IOException { LOG.info("Delete malformed table descriptor file {}", file); if (!fs.delete(file, false)) { LOG.warn("Failed to delete malformed table descriptor file {}", file); } } private static Optional> getTableDescriptorFromFs(FileSystem fs, Path tableDir, boolean readonly) throws IOException { Path tableInfoDir = new Path(tableDir, TABLEINFO_DIR); FileStatus[] descFiles = CommonFSUtils.listStatus(fs, tableInfoDir, TABLEINFO_PATHFILTER); if (descFiles == null || descFiles.length < 1) { return Optional.empty(); } Arrays.sort(descFiles, TABLEINFO_FILESTATUS_COMPARATOR); int i = 0; TableDescriptor td = null; FileStatus descFile = null; for (; i < descFiles.length; i++) { descFile = descFiles[i]; Path file = descFile.getPath(); // get file length from file name if present int fileLength = getTableInfoSequenceIdAndFileLength(file).fileLength; byte[] content = new byte[fileLength > 0 ? fileLength : Ints.checkedCast(descFile.getLen())]; try (FSDataInputStream in = fs.open(file)) { in.readFully(content); } catch (EOFException e) { LOG.info("Failed to load file {} due to EOF, it should be half written: {}", file, e.toString()); if (!readonly) { deleteMalformedFile(fs, file); } continue; } try { td = TableDescriptorBuilder.parseFrom(content); break; } catch (DeserializationException e) { LOG.info("Failed to parse file {} due to malformed protobuf message: {}", file, e.toString()); if (!readonly) { deleteMalformedFile(fs, file); } } } if (!readonly) { // i + 1 to skip the one we load for (i = i + 1; i < descFiles.length; i++) { Path file = descFiles[i].getPath(); LOG.info("Delete old table descriptor file {}", file); if (!fs.delete(file, false)) { LOG.info("Failed to delete old table descriptor file {}", file); } } } return td != null ? Optional.of(Pair.newPair(descFile, td)) : Optional.empty(); } @RestrictedApi(explanation = "Should only be called in tests", link = "", allowedOnPath = ".*/src/test/.*") public static void deleteTableDescriptors(FileSystem fs, Path tableDir) throws IOException { Path tableInfoDir = new Path(tableDir, TABLEINFO_DIR); deleteTableDescriptorFiles(fs, tableInfoDir, Integer.MAX_VALUE); } /** * Deletes files matching the table info file pattern within the given directory whose sequenceId * is at most the given max sequenceId. */ private static void deleteTableDescriptorFiles(FileSystem fs, Path dir, int maxSequenceId) throws IOException { FileStatus[] status = CommonFSUtils.listStatus(fs, dir, TABLEINFO_PATHFILTER); for (FileStatus file : status) { Path path = file.getPath(); int sequenceId = getTableInfoSequenceIdAndFileLength(path).sequenceId; if (sequenceId <= maxSequenceId) { boolean success = CommonFSUtils.delete(fs, path, false); if (success) { LOG.debug("Deleted {}", path); } else { LOG.error("Failed to delete table descriptor at {}", path); } } } } /** * Attempts to write a new table descriptor to the given table's directory. It begins at the * currentSequenceId + 1 and tries 10 times to find a new sequence number not already in use. *

* Removes the current descriptor file if passed in. * @return Descriptor file or null if we failed write. */ private static Path writeTableDescriptor(final FileSystem fs, final TableDescriptor td, final Path tableDir, final FileStatus currentDescriptorFile) throws IOException { // Here we will write to the final directory directly to avoid renaming as on OSS renaming is // not atomic and has performance issue. The reason why we could do this is that, in the below // code we will not overwrite existing files, we will write a new file instead. And when // loading, we will skip the half written file, please see the code in getTableDescriptorFromFs Path tableInfoDir = new Path(tableDir, TABLEINFO_DIR); // In proc v2 we have table lock so typically, there will be no concurrent writes. Keep the // retry logic here since we may still want to write the table descriptor from for example, // HBCK2? int currentSequenceId = currentDescriptorFile == null ? 0 : getTableInfoSequenceIdAndFileLength(currentDescriptorFile.getPath()).sequenceId; // Put arbitrary upperbound on how often we retry int maxAttempts = 10; int maxSequenceId = currentSequenceId + maxAttempts; byte[] bytes = TableDescriptorBuilder.toByteArray(td); for (int newSequenceId = currentSequenceId + 1; newSequenceId <= maxSequenceId; newSequenceId++) { String fileName = getTableInfoFileName(newSequenceId, bytes); Path filePath = new Path(tableInfoDir, fileName); try (FSDataOutputStream out = fs.create(filePath, false)) { out.write(bytes); } catch (FileAlreadyExistsException e) { LOG.debug("{} exists; retrying up to {} times", filePath, maxAttempts, e); continue; } catch (IOException e) { LOG.debug("Failed write {}; retrying up to {} times", filePath, maxAttempts, e); continue; } deleteTableDescriptorFiles(fs, tableInfoDir, newSequenceId - 1); return filePath; } return null; } /** * Create new TableDescriptor in HDFS. Happens when we are creating table. Used by tests. * @return True if we successfully created file. */ public boolean createTableDescriptor(TableDescriptor htd) throws IOException { return createTableDescriptor(htd, false); } /** * Create new TableDescriptor in HDFS. Happens when we are creating table. If forceCreation is * true then even if previous table descriptor is present it will be overwritten * @return True if we successfully created file. */ public boolean createTableDescriptor(TableDescriptor htd, boolean forceCreation) throws IOException { Path tableDir = getTableDir(htd.getTableName()); return createTableDescriptorForTableDirectory(tableDir, htd, forceCreation); } /** * Create a new TableDescriptor in HDFS in the specified table directory. Happens when we create a * new table during cluster start or in Clone and Create Table Procedures. Checks readOnly flag * passed on construction. * @param tableDir table directory under which we should write the file * @param htd description of the table to write * @param forceCreation if true,then even if previous table descriptor is present it will * be overwritten * @return true if the we successfully created the file, false if the file * already exists and we weren't forcing the descriptor creation. * @throws IOException if a filesystem error occurs */ public boolean createTableDescriptorForTableDirectory(Path tableDir, TableDescriptor htd, boolean forceCreation) throws IOException { if (this.fsreadonly) { throw new NotImplementedException("Cannot create a table descriptor - in read only mode"); } return createTableDescriptorForTableDirectory(this.fs, tableDir, htd, forceCreation); } /** * Create a new TableDescriptor in HDFS in the specified table directory. Happens when we create a * new table snapshoting. Does not enforce read-only. That is for caller to determine. * @param fs Filesystem to use. * @param tableDir table directory under which we should write the file * @param htd description of the table to write * @param forceCreation if true,then even if previous table descriptor is present it will * be overwritten * @return true if the we successfully created the file, false if the file * already exists and we weren't forcing the descriptor creation. * @throws IOException if a filesystem error occurs */ public static boolean createTableDescriptorForTableDirectory(FileSystem fs, Path tableDir, TableDescriptor htd, boolean forceCreation) throws IOException { Optional> opt = getTableDescriptorFromFs(fs, tableDir, false); if (opt.isPresent()) { LOG.debug("Current path={}", opt.get().getFirst()); if (!forceCreation) { if (htd.equals(opt.get().getSecond())) { LOG.trace("TableInfo already exists.. Skipping creation"); return false; } } } return writeTableDescriptor(fs, htd, tableDir, opt.map(Pair::getFirst).orElse(null)) != null; } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy