All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hudi.common.table.timeline.HoodieArchivedTimeline Maven / Gradle / Ivy

/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *      http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.hudi.common.table.timeline;

import org.apache.avro.generic.GenericRecord;
import org.apache.avro.generic.IndexedRecord;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.Path;
import org.apache.hudi.avro.model.HoodieArchivedMetaEntry;
import org.apache.hudi.common.model.HoodieLogFile;
import org.apache.hudi.common.model.HoodiePartitionMetadata;
import org.apache.hudi.common.table.HoodieTableMetaClient;
import org.apache.hudi.common.table.HoodieTimeline;
import org.apache.hudi.common.table.log.HoodieLogFormat;
import org.apache.hudi.common.table.log.block.HoodieAvroDataBlock;
import org.apache.hudi.common.util.Option;
import org.apache.hudi.exception.HoodieIOException;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;

import java.io.IOException;
import java.io.Serializable;
import java.nio.charset.StandardCharsets;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Comparator;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.function.Function;
import java.util.regex.Matcher;
import java.util.regex.Pattern;
import java.util.stream.Collectors;
import java.util.stream.Stream;

/**
 * Represents the Archived Timeline for the Hoodie table. Instants for the last 12 hours (configurable) is in the
 * ActiveTimeline and the rest are in ArchivedTimeline.
 * 

*

* Instants are read from the archive file during initialization and never refreshed. To refresh, clients need to call * reload() *

*

* This class can be serialized and de-serialized and on de-serialization the FileSystem is re-initialized. */ public class HoodieArchivedTimeline extends HoodieDefaultTimeline { private static final Pattern ARCHIVE_FILE_PATTERN = Pattern.compile("^\\.commits_\\.archive\\.([0-9]*)$"); private static final String HOODIE_COMMIT_ARCHIVE_LOG_FILE_PREFIX = "commits"; private static final String ACTION_TYPE_KEY = "actionType"; private HoodieTableMetaClient metaClient; private Map readCommits = new HashMap<>(); private static final Logger LOG = LogManager.getLogger(HoodieArchivedTimeline.class); /** * Loads instants between (startTs, endTs]. * Note that there is no lazy loading, so this may not work if really long time range (endTs-startTs) is specified. * TBD: Should we enforce maximum time range? */ public HoodieArchivedTimeline(HoodieTableMetaClient metaClient) { this.metaClient = metaClient; setInstants(this.loadInstants(false)); // multiple casts will make this lambda serializable - // http://docs.oracle.com/javase/specs/jls/se8/html/jls-15.html#jls-15.16 this.details = (Function> & Serializable) this::getInstantDetails; } /** * For serialization and de-serialization only. * * @deprecated */ public HoodieArchivedTimeline() {} /** * This method is only used when this object is deserialized in a spark executor. * * @deprecated */ private void readObject(java.io.ObjectInputStream in) throws IOException, ClassNotFoundException { in.defaultReadObject(); } public static Path getArchiveLogPath(String archiveFolder) { return new Path(archiveFolder, HOODIE_COMMIT_ARCHIVE_LOG_FILE_PREFIX); } public void loadInstantDetailsInMemory(String startTs, String endTs) { loadInstants(startTs, endTs); } public void clearInstantDetailsFromMemory(String startTs, String endTs) { this.findInstantsInRange(startTs, endTs).getInstants().forEach(instant -> this.readCommits.remove(instant.getTimestamp())); } @Override public Option getInstantDetails(HoodieInstant instant) { return Option.ofNullable(readCommits.get(instant.getTimestamp())); } public HoodieArchivedTimeline reload() { return new HoodieArchivedTimeline(metaClient); } private HoodieInstant readCommit(GenericRecord record, boolean loadDetails) { final String commitTime = record.get(HoodiePartitionMetadata.COMMIT_TIME_KEY).toString(); final String action = record.get(ACTION_TYPE_KEY).toString(); if (loadDetails) { Option.ofNullable(record.get(getMetadataKey(action))).map(actionData -> this.readCommits.put(commitTime, actionData.toString().getBytes(StandardCharsets.UTF_8)) ); } return new HoodieInstant(false, action, commitTime); } private String getMetadataKey(String action) { switch (action) { case HoodieTimeline.CLEAN_ACTION: return "hoodieCleanMetadata"; case HoodieTimeline.COMMIT_ACTION: return "hoodieCommitMetadata"; case HoodieTimeline.DELTA_COMMIT_ACTION: return "hoodieCommitMetadata"; case HoodieTimeline.ROLLBACK_ACTION: return "hoodieRollbackMetadata"; case HoodieTimeline.SAVEPOINT_ACTION: return "hoodieSavePointMetadata"; default: throw new HoodieIOException("Unknown action in metadata " + action); } } private List loadInstants(boolean loadInstantDetails) { return loadInstants(null, loadInstantDetails); } private List loadInstants(String startTs, String endTs) { return loadInstants(new TimeRangeFilter(startTs, endTs), true); } /** * This is method to read selected instants. Do NOT use this directly use one of the helper methods above * If loadInstantDetails is set to true, this would also update 'readCommits' map with commit details * If filter is specified, only the filtered instants are loaded */ private List loadInstants(TimeRangeFilter filter, boolean loadInstantDetails) { try { // list all files FileStatus[] fsStatuses = metaClient.getFs().globStatus( new Path(metaClient.getArchivePath() + "/.commits_.archive*")); // sort files by version suffix in reverse (implies reverse chronological order) Arrays.sort(fsStatuses, new ArchiveFileVersionComparator()); List instantsInRange = new ArrayList<>(); for (FileStatus fs : fsStatuses) { //read the archived file HoodieLogFormat.Reader reader = HoodieLogFormat.newReader(metaClient.getFs(), new HoodieLogFile(fs.getPath()), HoodieArchivedMetaEntry.getClassSchema()); try { int instantsInPreviousFile = instantsInRange.size(); //read the avro blocks while (reader.hasNext()) { HoodieAvroDataBlock blk = (HoodieAvroDataBlock) reader.next(); // TODO If we can store additional metadata in datablock, we can skip parsing records // (such as startTime, endTime of records in the block) List records = blk.getRecords(); // filter blocks in desired time window Stream instantsInBlkStream = records.stream() .map(r -> readCommit((GenericRecord) r, loadInstantDetails)); if (filter != null) { instantsInBlkStream = instantsInBlkStream.filter(filter::isInRange); } instantsInRange.addAll(instantsInBlkStream.collect(Collectors.toList())); } if (filter != null) { int instantsInCurrentFile = instantsInRange.size() - instantsInPreviousFile; if (instantsInPreviousFile > 0 && instantsInCurrentFile == 0) { // Note that this is an optimization to skip reading unnecessary archived files // This signals we crossed lower bound of desired time window. break; } } } finally { reader.close(); } } return instantsInRange; } catch (IOException e) { throw new HoodieIOException( "Could not load archived commit timeline from path " + metaClient.getArchivePath(), e); } } private static class TimeRangeFilter { private final String startTs; private final String endTs; public TimeRangeFilter(String startTs, String endTs) { this.startTs = startTs; this.endTs = endTs; } public boolean isInRange(HoodieInstant instant) { return HoodieTimeline.isInRange(instant.getTimestamp(), this.startTs, this.endTs); } } /** * Sort files by reverse order of version suffix in file name. */ public static class ArchiveFileVersionComparator implements Comparator, Serializable { @Override public int compare(FileStatus f1, FileStatus f2) { return Integer.compare(getArchivedFileSuffix(f2), getArchivedFileSuffix(f1)); } private int getArchivedFileSuffix(FileStatus f) { try { Matcher fileMatcher = ARCHIVE_FILE_PATTERN.matcher(f.getPath().getName()); if (fileMatcher.matches()) { return Integer.parseInt(fileMatcher.group(1)); } } catch (NumberFormatException e) { // log and ignore any format warnings LOG.warn("error getting suffix for archived file: " + f.getPath()); } // return default value in case of any errors return 0; } } }




© 2015 - 2025 Weber Informatics LLC | Privacy Policy