All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.flink.runtime.state.filesystem.FsStateBackend Maven / Gradle / Ivy

/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.flink.runtime.state.filesystem;

import org.apache.flink.annotation.PublicEvolving;
import org.apache.flink.api.common.JobID;
import org.apache.flink.api.common.typeutils.TypeSerializer;
import org.apache.flink.configuration.CheckpointingOptions;
import org.apache.flink.configuration.ReadableConfig;
import org.apache.flink.core.execution.SavepointFormatType;
import org.apache.flink.core.fs.CloseableRegistry;
import org.apache.flink.core.fs.FileSystem;
import org.apache.flink.core.fs.Path;
import org.apache.flink.metrics.MetricGroup;
import org.apache.flink.runtime.execution.Environment;
import org.apache.flink.runtime.query.TaskKvStateRegistry;
import org.apache.flink.runtime.state.AbstractKeyedStateBackend;
import org.apache.flink.runtime.state.AbstractStateBackend;
import org.apache.flink.runtime.state.BackendBuildingException;
import org.apache.flink.runtime.state.CheckpointStorageAccess;
import org.apache.flink.runtime.state.ConfigurableStateBackend;
import org.apache.flink.runtime.state.DefaultOperatorStateBackendBuilder;
import org.apache.flink.runtime.state.KeyGroupRange;
import org.apache.flink.runtime.state.KeyedStateHandle;
import org.apache.flink.runtime.state.LocalRecoveryConfig;
import org.apache.flink.runtime.state.OperatorStateBackend;
import org.apache.flink.runtime.state.OperatorStateHandle;
import org.apache.flink.runtime.state.TaskStateManager;
import org.apache.flink.runtime.state.heap.HeapKeyedStateBackendBuilder;
import org.apache.flink.runtime.state.heap.HeapPriorityQueueSetFactory;
import org.apache.flink.runtime.state.metrics.LatencyTrackingStateConfig;
import org.apache.flink.runtime.state.ttl.TtlTimeProvider;
import org.apache.flink.util.MathUtils;
import org.apache.flink.util.TernaryBoolean;

import org.slf4j.LoggerFactory;

import javax.annotation.Nonnull;
import javax.annotation.Nullable;

import java.io.IOException;
import java.net.URI;
import java.util.Collection;

import static org.apache.flink.configuration.CheckpointingOptions.FS_SMALL_FILE_THRESHOLD;
import static org.apache.flink.util.Preconditions.checkArgument;
import static org.apache.flink.util.Preconditions.checkNotNull;

/**
 * IMPORTANT {@link FsStateBackend} is deprecated in favor of {@link
 * org.apache.flink.runtime.state.hashmap.HashMapStateBackend} and {@link
 * org.apache.flink.runtime.state.storage.FileSystemCheckpointStorage}. This change does not affect
 * the runtime characteristics of your Jobs and is simply an API change to help better communicate
 * the ways Flink separates local state storage from fault tolerance. Jobs can be upgraded without
 * loss of state. If configuring your state backend via the {@code StreamExecutionEnvironment}
 * please make the following changes.
 *
 * 
{@code
 * 		StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
 * 		env.setStateBackend(new HashMapStateBackend());
 * 		env.getCheckpointConfig().setCheckpointStorage("hdfs:///checkpoints");
 * }
* *

If you are configuring your state backend via the {@code flink-conf.yaml} please make the * following changes set your state backend type to "hashmap" {@code state.backend: hashmap}. * *

This state backend holds the working state in the memory (JVM heap) of the TaskManagers. The * state backend checkpoints state as files to a file system (hence the backend's name). * *

Each checkpoint individually will store all its files in a subdirectory that includes the * checkpoint number, such as {@code hdfs://namenode:port/flink-checkpoints/chk-17/}. * *

State Size Considerations

* *

Working state is kept on the TaskManager heap. If a TaskManager executes multiple tasks * concurrently (if the TaskManager has multiple slots, or if slot-sharing is used) then the * aggregate state of all tasks needs to fit into that TaskManager's memory. * *

This state backend stores small state chunks directly with the metadata, to avoid creating * many small files. The threshold for that is configurable. When increasing this threshold, the * size of the checkpoint metadata increases. The checkpoint metadata of all retained completed * checkpoints needs to fit into the JobManager's heap memory. This is typically not a problem, * unless the threshold {@link #getMinFileSizeThreshold()} is increased significantly. * *

Persistence Guarantees

* *

Checkpoints from this state backend are as persistent and available as filesystem that is * written to. If the file system is a persistent distributed file system, this state backend * supports highly available setups. The backend additionally supports savepoints and externalized * checkpoints. * *

Configuration

* *

As for all state backends, this backend can either be configured within the application (by * creating the backend with the respective constructor parameters and setting it on the execution * environment) or by specifying it in the Flink configuration. * *

If the state backend was specified in the application, it may pick up additional configuration * parameters from the Flink configuration. For example, if the backend if configured in the * application without a default savepoint directory, it will pick up a default savepoint directory * specified in the Flink configuration of the running job/cluster. That behavior is implemented via * the {@link #configure(ReadableConfig, ClassLoader)} method. */ @Deprecated @PublicEvolving public class FsStateBackend extends AbstractFileStateBackend implements ConfigurableStateBackend { private static final long serialVersionUID = -8191916350224044011L; /** Maximum size of state that is stored with the metadata, rather than in files (1 MiByte). */ private static final int MAX_FILE_STATE_THRESHOLD = 1024 * 1024; // ------------------------------------------------------------------------ /** * State below this size will be stored as part of the metadata, rather than in files. A value * of '-1' means not yet configured, in which case the default will be used. */ private final int fileStateThreshold; /** * The write buffer size for created checkpoint stream, this should not be less than file state * threshold when we want state below that threshold stored as part of metadata not files. A * value of '-1' means not yet configured, in which case the default will be used. */ private final int writeBufferSize; // ----------------------------------------------------------------------- /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. */ public FsStateBackend(String checkpointDataUri) { this(new Path(checkpointDataUri)); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. * @param asynchronousSnapshots This parameter is only there for API compatibility. Checkpoints * are always asynchronous now. */ public FsStateBackend(String checkpointDataUri, boolean asynchronousSnapshots) { this(new Path(checkpointDataUri), asynchronousSnapshots); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. */ public FsStateBackend(Path checkpointDataUri) { this(checkpointDataUri.toUri()); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. * @param asynchronousSnapshots This parameter is only there for API compatibility. Checkpoints * are always asynchronous now. */ public FsStateBackend(Path checkpointDataUri, boolean asynchronousSnapshots) { this(checkpointDataUri.toUri(), asynchronousSnapshots); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. */ public FsStateBackend(URI checkpointDataUri) { this(checkpointDataUri, null, -1, -1, TernaryBoolean.UNDEFINED); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. Optionally, this constructor accepts a default savepoint storage * directory to which savepoints are stored when no custom target path is give to the savepoint * command. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. * @param defaultSavepointDirectory The default directory to store savepoints to. May be null. */ public FsStateBackend(URI checkpointDataUri, @Nullable URI defaultSavepointDirectory) { this(checkpointDataUri, defaultSavepointDirectory, -1, -1, TernaryBoolean.UNDEFINED); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. * @param asynchronousSnapshots This parameter is only there for API compatibility. Checkpoints * are always asynchronous now. */ public FsStateBackend(URI checkpointDataUri, boolean asynchronousSnapshots) { this(checkpointDataUri, null, -1, -1, TernaryBoolean.fromBoolean(asynchronousSnapshots)); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. * @param fileStateSizeThreshold State up to this size will be stored as part of the metadata, * rather than in files */ public FsStateBackend(URI checkpointDataUri, int fileStateSizeThreshold) { this(checkpointDataUri, null, fileStateSizeThreshold, -1, TernaryBoolean.UNDEFINED); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDataUri The URI describing the filesystem (scheme and optionally authority), * and the path to the checkpoint data directory. * @param fileStateSizeThreshold State up to this size will be stored as part of the metadata, * rather than in files (-1 for default value). * @param asynchronousSnapshots This parameter is only there for API compatibility. Checkpoints * are always asynchronous now. */ public FsStateBackend( URI checkpointDataUri, int fileStateSizeThreshold, boolean asynchronousSnapshots) { this( checkpointDataUri, null, fileStateSizeThreshold, -1, TernaryBoolean.fromBoolean(asynchronousSnapshots)); } /** * Creates a new state backend that stores its checkpoint data in the file system and location * defined by the given URI. * *

A file system for the file system scheme in the URI (e.g., 'file://', 'hdfs://', or * 'S3://') must be accessible via {@link FileSystem#get(URI)}. * *

For a state backend targeting HDFS, this means that the URI must either specify the * authority (host and port), or that the Hadoop configuration that describes that information * must be in the classpath. * * @param checkpointDirectory The path to write checkpoint metadata to. * @param defaultSavepointDirectory The path to write savepoints to. If null, the value from the * runtime configuration will be used, or savepoint target locations need to be passed when * triggering a savepoint. * @param fileStateSizeThreshold State below this size will be stored as part of the metadata, * rather than in files. If -1, the value configured in the runtime configuration will be * used, or the default value (1KB) if nothing is configured. * @param writeBufferSize Write buffer size used to serialize state. If -1, the value configured * in the runtime configuration will be used, or the default value (4KB) if nothing is * configured. * @param asynchronousSnapshots This parameter is only there for API compatibility. Checkpoints * are always asynchronous now. */ public FsStateBackend( URI checkpointDirectory, @Nullable URI defaultSavepointDirectory, int fileStateSizeThreshold, int writeBufferSize, @SuppressWarnings("unused") TernaryBoolean asynchronousSnapshots) { super( checkNotNull(checkpointDirectory, "checkpoint directory is null"), defaultSavepointDirectory); checkArgument( fileStateSizeThreshold >= -1 && fileStateSizeThreshold <= MAX_FILE_STATE_THRESHOLD, "The threshold for file state size must be in [-1, %s], where '-1' means to use " + "the value from the deployment's configuration.", MAX_FILE_STATE_THRESHOLD); checkArgument( writeBufferSize >= -1, "The write buffer size must be not less than '-1', where '-1' means to use " + "the value from the deployment's configuration."); this.fileStateThreshold = fileStateSizeThreshold; this.writeBufferSize = writeBufferSize; } /** * Private constructor that creates a re-configured copy of the state backend. * * @param original The state backend to re-configure * @param configuration The configuration */ private FsStateBackend( FsStateBackend original, ReadableConfig configuration, ClassLoader classLoader) { super(original.getCheckpointPath(), original.getSavepointPath(), configuration); if (getValidFileStateThreshold(original.fileStateThreshold) >= 0) { this.fileStateThreshold = original.fileStateThreshold; } else { final int configuredStateThreshold = getValidFileStateThreshold( configuration.get(FS_SMALL_FILE_THRESHOLD).getBytes()); if (configuredStateThreshold >= 0) { this.fileStateThreshold = configuredStateThreshold; } else { this.fileStateThreshold = MathUtils.checkedDownCast( FS_SMALL_FILE_THRESHOLD.defaultValue().getBytes()); // because this is the only place we (unlikely) ever log, we lazily // create the logger here LoggerFactory.getLogger(AbstractFileStateBackend.class) .warn( "Ignoring invalid file size threshold value ({}): {} - using default value {} instead.", FS_SMALL_FILE_THRESHOLD.key(), configuration.get(FS_SMALL_FILE_THRESHOLD).getBytes(), FS_SMALL_FILE_THRESHOLD.defaultValue()); } } final int bufferSize = original.writeBufferSize >= 0 ? original.writeBufferSize : configuration.get(CheckpointingOptions.FS_WRITE_BUFFER_SIZE); this.writeBufferSize = Math.max(bufferSize, this.fileStateThreshold); // configure latency tracking latencyTrackingConfigBuilder = original.latencyTrackingConfigBuilder.configure(configuration); } private int getValidFileStateThreshold(long fileStateThreshold) { if (fileStateThreshold >= 0 && fileStateThreshold <= MAX_FILE_STATE_THRESHOLD) { return (int) fileStateThreshold; } return -1; } // ------------------------------------------------------------------------ // Properties // ------------------------------------------------------------------------ /** * Gets the base directory where all the checkpoints are stored. The job-specific checkpoint * directory is created inside this directory. * * @return The base directory for checkpoints. * @deprecated Deprecated in favor of {@link #getCheckpointPath()}. */ @Deprecated public Path getBasePath() { return getCheckpointPath(); } /** * Gets the base directory where all the checkpoints are stored. The job-specific checkpoint * directory is created inside this directory. * * @return The base directory for checkpoints. */ @Nonnull @Override public Path getCheckpointPath() { // we know that this can never be null by the way of constructor checks //noinspection ConstantConditions return super.getCheckpointPath(); } /** * Gets the threshold below which state is stored as part of the metadata, rather than in files. * This threshold ensures that the backend does not create a large amount of very small files, * where potentially the file pointers are larger than the state itself. * *

If not explicitly configured, this is the default value of {@link * CheckpointingOptions#FS_SMALL_FILE_THRESHOLD}. * * @return The file size threshold, in bytes. */ public int getMinFileSizeThreshold() { return fileStateThreshold >= 0 ? fileStateThreshold : MathUtils.checkedDownCast(FS_SMALL_FILE_THRESHOLD.defaultValue().getBytes()); } /** * Gets the write buffer size for created checkpoint stream. * *

If not explicitly configured, this is the default value of {@link * CheckpointingOptions#FS_WRITE_BUFFER_SIZE}. * * @return The write buffer size, in bytes. */ public int getWriteBufferSize() { return writeBufferSize >= 0 ? writeBufferSize : CheckpointingOptions.FS_WRITE_BUFFER_SIZE.defaultValue(); } /** * Gets whether the key/value data structures are asynchronously snapshotted, which is always * true for this state backend. */ public boolean isUsingAsynchronousSnapshots() { return true; } @Override public boolean supportsNoClaimRestoreMode() { // we never share any files, all snapshots are full return true; } @Override public boolean supportsSavepointFormat(SavepointFormatType formatType) { return true; } // ------------------------------------------------------------------------ // Reconfiguration // ------------------------------------------------------------------------ /** * Creates a copy of this state backend that uses the values defined in the configuration for * fields where that were not specified in this state backend. * * @param config the configuration * @return The re-configured variant of the state backend */ @Override public FsStateBackend configure(ReadableConfig config, ClassLoader classLoader) { return new FsStateBackend(this, config, classLoader); } // ------------------------------------------------------------------------ // initialization and cleanup // ------------------------------------------------------------------------ @Override public CheckpointStorageAccess createCheckpointStorage(JobID jobId) throws IOException { checkNotNull(jobId, "jobId"); return new FsCheckpointStorageAccess( getCheckpointPath(), getSavepointPath(), jobId, getMinFileSizeThreshold(), getWriteBufferSize()); } // ------------------------------------------------------------------------ // state holding structures // ------------------------------------------------------------------------ @Override public AbstractKeyedStateBackend createKeyedStateBackend( Environment env, JobID jobID, String operatorIdentifier, TypeSerializer keySerializer, int numberOfKeyGroups, KeyGroupRange keyGroupRange, TaskKvStateRegistry kvStateRegistry, TtlTimeProvider ttlTimeProvider, MetricGroup metricGroup, @Nonnull Collection stateHandles, CloseableRegistry cancelStreamRegistry) throws BackendBuildingException { TaskStateManager taskStateManager = env.getTaskStateManager(); LocalRecoveryConfig localRecoveryConfig = taskStateManager.createLocalRecoveryConfig(); HeapPriorityQueueSetFactory priorityQueueSetFactory = new HeapPriorityQueueSetFactory(keyGroupRange, numberOfKeyGroups, 128); LatencyTrackingStateConfig latencyTrackingStateConfig = latencyTrackingConfigBuilder.setMetricGroup(metricGroup).build(); return new HeapKeyedStateBackendBuilder<>( kvStateRegistry, keySerializer, env.getUserCodeClassLoader().asClassLoader(), numberOfKeyGroups, keyGroupRange, env.getExecutionConfig(), ttlTimeProvider, latencyTrackingStateConfig, stateHandles, AbstractStateBackend.getCompressionDecorator(env.getExecutionConfig()), localRecoveryConfig, priorityQueueSetFactory, isUsingAsynchronousSnapshots(), cancelStreamRegistry) .build(); } @Override public OperatorStateBackend createOperatorStateBackend( Environment env, String operatorIdentifier, @Nonnull Collection stateHandles, CloseableRegistry cancelStreamRegistry) throws BackendBuildingException { return new DefaultOperatorStateBackendBuilder( env.getUserCodeClassLoader().asClassLoader(), env.getExecutionConfig(), isUsingAsynchronousSnapshots(), stateHandles, cancelStreamRegistry) .build(); } // ------------------------------------------------------------------------ // utilities // ------------------------------------------------------------------------ @Override public String toString() { return "File State Backend (" + "checkpoints: '" + getCheckpointPath() + "', savepoints: '" + getSavepointPath() + ", fileStateThreshold: " + fileStateThreshold + ")"; } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy