com.google.cloud.bigtable.hbase2_x.replication.HbaseToCloudBigtableReplicationEndpoint Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of bigtable-hbase-2.x-replication Show documentation
Show all versions of bigtable-hbase-2.x-replication Show documentation
Library to enable one way replication from HBase to Cloud Bigtable.
/*
* Copyright 2022 Google LLC
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.google.cloud.bigtable.hbase2_x.replication;
// TODO reverse the package names for 1.x and 2.x to com.google.cloud.bigtable.replication.hbase1_x
import static org.apache.hadoop.hbase.replication.BaseReplicationEndpoint.REPLICATION_WALENTRYFILTER_CONFIG_KEY;
import com.google.bigtable.repackaged.com.google.api.core.InternalExtensionOnly;
import com.google.bigtable.repackaged.com.google.common.util.concurrent.AbstractService;
import com.google.cloud.bigtable.hbase.replication.CloudBigtableReplicator;
import com.google.cloud.bigtable.hbase.replication.adapters.BigtableWALEntry;
import com.google.cloud.bigtable.hbase2_x.replication.metrics.HBaseMetricsExporter;
import java.io.IOException;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.UUID;
import org.apache.hadoop.hbase.replication.ChainWALEntryFilter;
import org.apache.hadoop.hbase.replication.NamespaceTableCfWALEntryFilter;
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
import org.apache.hadoop.hbase.replication.ReplicationPeer;
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
import org.apache.hadoop.hbase.replication.ScopeWALEntryFilter;
import org.apache.hadoop.hbase.replication.WALEntryFilter;
import org.apache.hadoop.hbase.wal.WAL;
import org.apache.hbase.thirdparty.com.google.common.collect.Lists;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
@InternalExtensionOnly
public class HbaseToCloudBigtableReplicationEndpoint extends AbstractService
implements ReplicationEndpoint {
private final CloudBigtableReplicator cloudBigtableReplicator;
private final HBaseMetricsExporter metricsExporter;
private Context context;
private static final Logger LOG =
LoggerFactory.getLogger(HbaseToCloudBigtableReplicationEndpoint.class);
public HbaseToCloudBigtableReplicationEndpoint() {
cloudBigtableReplicator = new CloudBigtableReplicator();
metricsExporter = new HBaseMetricsExporter();
}
// User of the class should use this class after calling init.
// https://github.com/apache/hbase/blob/master/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java#L321
@Override
public void init(Context context) throws IOException {
this.context = context;
if (this.context != null) {
ReplicationPeer peer = this.context.getReplicationPeer();
if (peer != null) {
peer.registerPeerConfigListener(this);
} else {
LOG.warn(
"Not tracking replication peer config changes for Peer Id "
+ this.context.getPeerId()
+ " because there's no such peer");
}
}
}
@Override
public boolean canReplicateToSameCluster() {
return false;
}
@Override
public UUID getPeerUUID() {
return cloudBigtableReplicator.getPeerUUID();
}
// the default filter excludes the system tables and already replicated table
// https://github.com/apache/hbase/blob/master/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java#L330
// Keeping the implementation same as HBaseReplicationEndpoint
@Override
public WALEntryFilter getWALEntryfilter() {
ArrayList filters = Lists.newArrayList();
WALEntryFilter scopeFilter = getScopeWALEntryFilter();
if (scopeFilter != null) {
filters.add(scopeFilter);
}
WALEntryFilter tableCfFilter = getNamespaceTableCfWALEntryFilter();
if (tableCfFilter != null) {
filters.add(tableCfFilter);
}
if (context != null && context.getPeerConfig() != null) {
String filterNameCSV =
context.getPeerConfig().getConfiguration().get(REPLICATION_WALENTRYFILTER_CONFIG_KEY);
if (filterNameCSV != null && !filterNameCSV.isEmpty()) {
String[] filterNames = filterNameCSV.split(",");
for (String filterName : filterNames) {
try {
Class> clazz = Class.forName(filterName);
filters.add((WALEntryFilter) clazz.getDeclaredConstructor().newInstance());
} catch (Exception e) {
LOG.error("Unable to create WALEntryFilter " + filterName, e);
}
}
}
}
return filters.isEmpty() ? null : new ChainWALEntryFilter(filters);
}
/**
* Returns a WALEntryFilter for checking the scope. Subclasses can return null if they don't want
* this filter
*/
private WALEntryFilter getScopeWALEntryFilter() {
return new ScopeWALEntryFilter();
}
/**
* Returns a WALEntryFilter for checking replication per table and CF. Subclasses can return null
* if they don't want this filter
*/
private WALEntryFilter getNamespaceTableCfWALEntryFilter() {
return new NamespaceTableCfWALEntryFilter(context.getReplicationPeer());
}
@Override
public boolean replicate(ReplicateContext replicateContext) {
Map> walEntriesByTable = new HashMap<>();
for (WAL.Entry wal : replicateContext.getEntries()) {
String tableName = wal.getKey().getTableName().getNameAsString();
BigtableWALEntry bigtableWALEntry =
new BigtableWALEntry(wal.getKey().getWriteTime(), wal.getEdit().getCells(), tableName);
if (!walEntriesByTable.containsKey(tableName)) {
walEntriesByTable.put(tableName, new ArrayList<>());
}
walEntriesByTable.get(tableName).add(bigtableWALEntry);
}
return cloudBigtableReplicator.replicate(walEntriesByTable);
}
@Override
public boolean isStarting() {
return state() == State.STARTING;
}
@Override
public void start() {
startAsync();
}
@Override
public void stop() {
stopAsync();
}
@Override
protected void doStart() {
metricsExporter.setMetricsSource(context.getMetrics());
cloudBigtableReplicator.start(context.getConfiguration(), metricsExporter);
notifyStarted();
}
@Override
protected void doStop() {
cloudBigtableReplicator.stop();
notifyStopped();
}
// It is a no op for HBaseReplicationEndpoint. Keeping as same.
@Override
public void peerConfigUpdated(ReplicationPeerConfig replicationPeerConfig) {}
// TODO(we can implement this to enable/disable dry-run mode without deleting the peer)
}