All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.hbase.regionserver.SplitLogWorker Maven / Gradle / Ivy

There is a newer version: 3.0.0-beta-1
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hbase.regionserver;

import java.io.FileNotFoundException;
import java.io.IOException;
import java.io.InterruptedIOException;
import java.net.ConnectException;
import java.net.SocketTimeoutException;
import java.util.Optional;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.NotServingRegionException;
import org.apache.hadoop.hbase.Server;
import org.apache.hadoop.hbase.client.RetriesExhaustedException;
import org.apache.hadoop.hbase.coordination.SplitLogWorkerCoordination;
import org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status;
import org.apache.hadoop.hbase.replication.ReplicationPeerImpl;
import org.apache.hadoop.hbase.replication.ReplicationUtils;
import org.apache.hadoop.hbase.replication.SyncReplicationState;
import org.apache.hadoop.hbase.util.CancelableProgressable;
import org.apache.hadoop.hbase.util.CommonFSUtils;
import org.apache.hadoop.hbase.util.ExceptionUtil;
import org.apache.hadoop.hbase.util.FSUtils;
import org.apache.hadoop.hbase.util.Pair;
import org.apache.hadoop.hbase.wal.SyncReplicationWALProvider;
import org.apache.hadoop.hbase.wal.WALFactory;
import org.apache.hadoop.hbase.wal.WALSplitter;
import org.apache.yetus.audience.InterfaceAudience;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

/**
 * This worker is spawned in every regionserver, including master. The Worker waits for log
 * splitting tasks to be put up by the {@link org.apache.hadoop.hbase.master.SplitLogManager}
 * running in the master and races with other workers in other serves to acquire those tasks. The
 * coordination is done via coordination engine.
 * 

* If a worker has successfully moved the task from state UNASSIGNED to OWNED then it owns the task. * It keeps heart beating the manager by periodically moving the task from UNASSIGNED to OWNED * state. On success it moves the task to TASK_DONE. On unrecoverable error it moves task state to * ERR. If it cannot continue but wants the master to retry the task then it moves the task state to * RESIGNED. *

* The manager can take a task away from a worker by moving the task from OWNED to UNASSIGNED. In * the absence of a global lock there is a unavoidable race here - a worker might have just finished * its task when it is stripped of its ownership. Here we rely on the idempotency of the log * splitting task for correctness * @deprecated since 2.4.0 and in 3.0.0, to be removed in 4.0.0, replaced by procedure-based * distributed WAL splitter, see SplitWALRemoteProcedure */ @Deprecated @InterfaceAudience.Private public class SplitLogWorker implements Runnable { private static final Logger LOG = LoggerFactory.getLogger(SplitLogWorker.class); Thread worker; // thread pool which executes recovery work private final SplitLogWorkerCoordination coordination; private final RegionServerServices server; public SplitLogWorker(Server hserver, Configuration conf, RegionServerServices server, TaskExecutor splitTaskExecutor) { this.server = server; this.coordination = hserver.getCoordinatedStateManager().getSplitLogWorkerCoordination(); coordination.init(server, conf, splitTaskExecutor, this); } public SplitLogWorker(Configuration conf, RegionServerServices server, LastSequenceId sequenceIdChecker, WALFactory factory) { this(server, conf, server, (f, p) -> splitLog(f, p, conf, server, sequenceIdChecker, factory)); } // returns whether we need to continue the split work private static boolean processSyncReplicationWAL(String name, Configuration conf, RegionServerServices server, FileSystem fs, Path walDir) throws IOException { Path walFile = new Path(walDir, name); String filename = walFile.getName(); Optional optSyncPeerId = SyncReplicationWALProvider.getSyncReplicationPeerIdFromWALName(filename); if (!optSyncPeerId.isPresent()) { return true; } String peerId = optSyncPeerId.get(); ReplicationPeerImpl peer = server.getReplicationSourceService().getReplicationPeers().getPeer(peerId); if (peer == null || !peer.getPeerConfig().isSyncReplication()) { return true; } Pair stateAndNewState = peer.getSyncReplicationStateAndNewState(); if ( stateAndNewState.getFirst().equals(SyncReplicationState.ACTIVE) && stateAndNewState.getSecond().equals(SyncReplicationState.NONE) ) { // copy the file to remote and overwrite the previous one String remoteWALDir = peer.getPeerConfig().getRemoteWALDir(); Path remoteWALDirForPeer = ReplicationUtils.getPeerRemoteWALDir(remoteWALDir, peerId); Path tmpRemoteWAL = new Path(remoteWALDirForPeer, filename + ".tmp"); FileSystem remoteFs = ReplicationUtils.getRemoteWALFileSystem(conf, remoteWALDir); try (FSDataInputStream in = fs.open(walFile); FSDataOutputStream out = remoteFs.createNonRecursive(tmpRemoteWAL, true, CommonFSUtils.getDefaultBufferSize(remoteFs), remoteFs.getDefaultReplication(tmpRemoteWAL), remoteFs.getDefaultBlockSize(tmpRemoteWAL), null)) { IOUtils.copy(in, out); } Path toCommitRemoteWAL = new Path(remoteWALDirForPeer, filename + ReplicationUtils.RENAME_WAL_SUFFIX); // Some FileSystem implementations may not support atomic rename so we need to do it in two // phases FSUtils.renameFile(remoteFs, tmpRemoteWAL, toCommitRemoteWAL); FSUtils.renameFile(remoteFs, toCommitRemoteWAL, new Path(remoteWALDirForPeer, filename)); } else if ( (stateAndNewState.getFirst().equals(SyncReplicationState.ACTIVE) && stateAndNewState.getSecond().equals(SyncReplicationState.STANDBY)) || stateAndNewState.getFirst().equals(SyncReplicationState.STANDBY) ) { // check whether we still need to process this file // actually we only write wal file which name is ended with .syncrep in A state, and after // transiting to a state other than A, we will reopen all the regions so the data in the wal // will be flushed so the wal file will be archived soon. But it is still possible that there // is a server crash when we are transiting from A to S, to simplify the logic of the transit // procedure, here we will also check the remote snapshot directory in state S, so that we do // not need wait until all the wal files with .syncrep suffix to be archived before finishing // the procedure. String remoteWALDir = peer.getPeerConfig().getRemoteWALDir(); Path remoteSnapshotDirForPeer = ReplicationUtils.getPeerSnapshotWALDir(remoteWALDir, peerId); FileSystem remoteFs = ReplicationUtils.getRemoteWALFileSystem(conf, remoteWALDir); if (remoteFs.exists(new Path(remoteSnapshotDirForPeer, filename))) { // the file has been replayed when the remote cluster was transited from S to DA, the // content will be replicated back to us so give up split it. LOG.warn("Giveup splitting {} since it has been replayed in the remote cluster and " + "the content will be replicated back", filename); return false; } } return true; } /** Returns Result either DONE, RESIGNED, or ERR. */ static Status splitLog(String filename, CancelableProgressable p, Configuration conf, RegionServerServices server, LastSequenceId sequenceIdChecker, WALFactory factory) { Path walDir; FileSystem fs; try { walDir = CommonFSUtils.getWALRootDir(conf); fs = walDir.getFileSystem(conf); } catch (IOException e) { LOG.warn("Resigning, could not find root dir or fs", e); return Status.RESIGNED; } try { if (!processSyncReplicationWAL(filename, conf, server, fs, walDir)) { return Status.DONE; } } catch (IOException e) { LOG.warn("failed to process sync replication wal {}", filename, e); return Status.RESIGNED; } // TODO have to correctly figure out when log splitting has been // interrupted or has encountered a transient error and when it has // encountered a bad non-retry-able persistent error. try { SplitLogWorkerCoordination splitLogWorkerCoordination = server.getCoordinatedStateManager() == null ? null : server.getCoordinatedStateManager().getSplitLogWorkerCoordination(); if ( !WALSplitter.splitLogFile(walDir, fs.getFileStatus(new Path(walDir, filename)), fs, conf, p, sequenceIdChecker, splitLogWorkerCoordination, factory, server) ) { return Status.PREEMPTED; } } catch (InterruptedIOException iioe) { LOG.warn("Resigning, interrupted splitting WAL {}", filename, iioe); return Status.RESIGNED; } catch (IOException e) { if (e instanceof FileNotFoundException) { // A wal file may not exist anymore. Nothing can be recovered so move on LOG.warn("Done, WAL {} does not exist anymore", filename, e); return Status.DONE; } Throwable cause = e.getCause(); if ( e instanceof RetriesExhaustedException && (cause instanceof NotServingRegionException || cause instanceof ConnectException || cause instanceof SocketTimeoutException) ) { LOG.warn("Resigning, can't connect to target regionserver splitting WAL {}", filename, e); return Status.RESIGNED; } else if (cause instanceof InterruptedException) { LOG.warn("Resigning, interrupted splitting WAL {}", filename, e); return Status.RESIGNED; } LOG.warn("Error splitting WAL {}", filename, e); return Status.ERR; } LOG.debug("Done splitting WAL {}", filename); return Status.DONE; } @Override public void run() { try { LOG.info("SplitLogWorker " + server.getServerName() + " starting"); coordination.registerListener(); // wait for Coordination Engine is ready boolean res = false; while (!res && !coordination.isStop()) { res = coordination.isReady(); } if (!coordination.isStop()) { coordination.taskLoop(); } } catch (Throwable t) { if (ExceptionUtil.isInterrupt(t)) { LOG.info("SplitLogWorker interrupted. Exiting. " + (coordination.isStop() ? "" : " (ERROR: exitWorker is not set, exiting anyway)")); } else { // only a logical error can cause here. Printing it out // to make debugging easier LOG.error("unexpected error ", t); } } finally { coordination.removeListener(); LOG.info("SplitLogWorker " + server.getServerName() + " exiting"); } } /** * If the worker is doing a task i.e. splitting a log file then stop the task. It doesn't exit the * worker thread. */ public void stopTask() { LOG.info("Sending interrupt to stop the worker thread"); worker.interrupt(); // TODO interrupt often gets swallowed, do what else? } /** * start the SplitLogWorker thread */ public void start() { worker = new Thread(null, this, "SplitLogWorker-" + server.getServerName().toShortString()); worker.start(); } /** * stop the SplitLogWorker thread */ public void stop() { coordination.stopProcessingTasks(); stopTask(); } /** * Objects implementing this interface actually do the task that has been acquired by a * {@link SplitLogWorker}. Since there isn't a water-tight guarantee that two workers will not be * executing the same task therefore it is better to have workers prepare the task and then have * the {@link org.apache.hadoop.hbase.master.SplitLogManager} commit the work in * SplitLogManager.TaskFinisher */ @FunctionalInterface public interface TaskExecutor { enum Status { DONE(), ERR(), RESIGNED(), PREEMPTED() } Status exec(String name, CancelableProgressable p); } /** * Returns the number of tasks processed by coordination. This method is used by tests only */ public int getTaskReadySeq() { return coordination.getTaskReadySeq(); } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy