All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.hbase.client.HBaseHbck Maven / Gradle / Ivy

There is a newer version: 3.0.0-beta-1
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hbase.client;

import java.io.IOException;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.concurrent.Callable;
import java.util.stream.Collectors;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.ServerName;
import org.apache.hadoop.hbase.ipc.RpcControllerFactory;
import org.apache.hadoop.hbase.master.RegionState;
import org.apache.yetus.audience.InterfaceAudience;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import org.apache.hbase.thirdparty.com.google.protobuf.ServiceException;

import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.HbckService.BlockingInterface;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse;
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse;

/**
 * Use {@link ClusterConnection#getHbck()} to obtain an instance of {@link Hbck} instead of
 * constructing an HBaseHbck directly.
 * 

* Connection should be an unmanaged connection obtained via * {@link ConnectionFactory#createConnection(Configuration)}. *

*

* NOTE: The methods in here can do damage to a cluster if applied in the wrong sequence or at the * wrong time. Use with caution. For experts only. These methods are only for the extreme case where * the cluster has been damaged or has achieved an inconsistent state because of some unforeseen * circumstance or bug and requires manual intervention. *

* An instance of this class is lightweight and not-thread safe. A new instance should be created by * each thread. Pooling or caching of the instance is not recommended. *

* @see ConnectionFactory * @see ClusterConnection * @see Hbck */ @InterfaceAudience.Private public class HBaseHbck implements Hbck { private static final Logger LOG = LoggerFactory.getLogger(HBaseHbck.class); private boolean aborted; private final BlockingInterface hbck; private RpcControllerFactory rpcControllerFactory; HBaseHbck(BlockingInterface hbck, RpcControllerFactory rpcControllerFactory) { this.hbck = hbck; this.rpcControllerFactory = rpcControllerFactory; } @Override public void close() throws IOException { // currently does nothing } @Override public void abort(String why, Throwable e) { this.aborted = true; // Currently does nothing but throw the passed message and exception throw new RuntimeException(why, e); } @Override public boolean isAborted() { return this.aborted; } @Override public TableState setTableStateInMeta(TableState state) throws IOException { try { GetTableStateResponse response = hbck.setTableStateInMeta(rpcControllerFactory.newController(), RequestConverter.buildSetTableStateInMetaRequest(state)); return TableState.convert(state.getTableName(), response.getTableState()); } catch (ServiceException se) { LOG.debug("table={}, state={}", state.getTableName(), state.getState(), se); throw new IOException(se); } } @Override public Map setRegionStateInMeta( Map nameOrEncodedName2State) throws IOException { try { if (LOG.isDebugEnabled()) { nameOrEncodedName2State.forEach((k, v) -> LOG.debug("region={}, state={}", k, v)); } MasterProtos.SetRegionStateInMetaResponse response = hbck.setRegionStateInMeta(rpcControllerFactory.newController(), RequestConverter.buildSetRegionStateInMetaRequest(nameOrEncodedName2State)); Map result = new HashMap<>(); for (RegionSpecifierAndState nameAndState : response.getStatesList()) { result.put(nameAndState.getRegionSpecifier().getValue().toStringUtf8(), RegionState.State.convert(nameAndState.getState())); } return result; } catch (ServiceException se) { throw new IOException(se); } } @Override public List assigns(List encodedRegionNames, boolean override) throws IOException { try { AssignsResponse response = this.hbck.assigns(rpcControllerFactory.newController(), RequestConverter.toAssignRegionsRequest(encodedRegionNames, override)); return response.getPidList(); } catch (ServiceException se) { LOG.debug(toCommaDelimitedString(encodedRegionNames), se); throw new IOException(se); } } @Override public List unassigns(List encodedRegionNames, boolean override) throws IOException { try { UnassignsResponse response = this.hbck.unassigns(rpcControllerFactory.newController(), RequestConverter.toUnassignRegionsRequest(encodedRegionNames, override)); return response.getPidList(); } catch (ServiceException se) { LOG.debug(toCommaDelimitedString(encodedRegionNames), se); throw new IOException(se); } } private static String toCommaDelimitedString(List list) { return list.stream().collect(Collectors.joining(", ")); } @Override public List bypassProcedure(List pids, long waitTime, boolean override, boolean recursive) throws IOException { BypassProcedureResponse response = ProtobufUtil.call(new Callable() { @Override public BypassProcedureResponse call() throws Exception { try { return hbck.bypassProcedure(rpcControllerFactory.newController(), BypassProcedureRequest.newBuilder().addAllProcId(pids).setWaitTime(waitTime) .setOverride(override).setRecursive(recursive).build()); } catch (Throwable t) { LOG.error(pids.stream().map(i -> i.toString()).collect(Collectors.joining(", ")), t); throw t; } } }); return response.getBypassedList(); } @Override public List scheduleServerCrashProcedures(List serverNames) throws IOException { try { ScheduleServerCrashProcedureResponse response = this.hbck.scheduleServerCrashProcedure(rpcControllerFactory.newController(), RequestConverter.toScheduleServerCrashProcedureRequest(serverNames)); return response.getPidList(); } catch (ServiceException se) { LOG.debug(toCommaDelimitedString( serverNames.stream().map(serverName -> ProtobufUtil.toServerName(serverName).toString()) .collect(Collectors.toList())), se); throw new IOException(se); } } @Override public List scheduleSCPsForUnknownServers() throws IOException { try { ScheduleSCPsForUnknownServersResponse response = this.hbck.scheduleSCPsForUnknownServers(rpcControllerFactory.newController(), ScheduleSCPsForUnknownServersRequest.newBuilder().build()); return response.getPidList(); } catch (ServiceException se) { LOG.debug("Failed to run ServerCrashProcedures for unknown servers", se); throw new IOException(se); } } @Override public boolean runHbckChore() throws IOException { try { RunHbckChoreResponse response = this.hbck.runHbckChore(rpcControllerFactory.newController(), RunHbckChoreRequest.newBuilder().build()); return response.getRan(); } catch (ServiceException se) { LOG.debug("Failed to run HBCK chore", se); throw new IOException(se); } } @Override public void fixMeta() throws IOException { try { this.hbck.fixMeta(rpcControllerFactory.newController(), FixMetaRequest.newBuilder().build()); } catch (ServiceException se) { throw new IOException(se); } } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy