org.apache.hadoop.hbase.backup.BackupInfo Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of hbase-server Show documentation
Show all versions of hbase-server Show documentation
Server functionality for HBase
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hbase.backup;
import java.io.IOException;
import java.io.InputStream;
import java.util.ArrayList;
import java.util.Calendar;
import java.util.Date;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.Set;
import org.apache.commons.lang.StringUtils;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.backup.util.BackupUtils;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos;
import org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupInfo.Builder;
import org.apache.hadoop.hbase.util.Bytes;
/**
* An object to encapsulate the information for each backup session
*/
@InterfaceAudience.Private
public class BackupInfo implements Comparable {
private static final Log LOG = LogFactory.getLog(BackupInfo.class);
public static interface Filter {
/**
* Filter interface
* @param info backup info
* @return true if info passes filter, false otherwise
*/
public boolean apply(BackupInfo info);
}
/**
* Backup session states
*/
public static enum BackupState {
RUNNING, COMPLETE, FAILED, ANY;
}
/**
* BackupPhase - phases of an ACTIVE backup session (running), when state of a backup session is
* BackupState.RUNNING
*/
public static enum BackupPhase {
REQUEST, SNAPSHOT, PREPARE_INCREMENTAL, SNAPSHOTCOPY, INCREMENTAL_COPY, STORE_MANIFEST;
}
/**
* Backup id
*/
private String backupId;
/**
* Backup type, full or incremental
*/
private BackupType type;
/**
* Target root directory for storing the backup files
*/
private String backupRootDir;
/**
* Backup state
*/
private BackupState state;
/**
* Backup phase
*/
private BackupPhase phase = BackupPhase.REQUEST;
/**
* Backup failure message
*/
private String failedMsg;
/**
* Backup status map for all tables
*/
private Map backupTableInfoMap;
/**
* Actual start timestamp of a backup process
*/
private long startTs;
/**
* Actual end timestamp of the backup process
*/
private long completeTs;
/**
* Total bytes of incremental logs copied
*/
private long totalBytesCopied;
/**
* For incremental backup, a location of a backed-up hlogs
*/
private String hlogTargetDir = null;
/**
* Incremental backup file list
*/
private List incrBackupFileList;
/**
* New region server log timestamps for table set after distributed log roll key - table name,
* value - map of RegionServer hostname -> last log rolled timestamp
*/
private HashMap> tableSetTimestampMap;
/**
* Backup progress in %% (0-100)
*/
private int progress;
/**
* Number of parallel workers. -1 - system defined
*/
private int workers = -1;
/**
* Bandwidth per worker in MB per sec. -1 - unlimited
*/
private long bandwidth = -1;
public BackupInfo() {
backupTableInfoMap = new HashMap();
}
public BackupInfo(String backupId, BackupType type, TableName[] tables, String targetRootDir) {
this();
this.backupId = backupId;
this.type = type;
this.backupRootDir = targetRootDir;
this.addTables(tables);
if (type == BackupType.INCREMENTAL) {
setHLogTargetDir(BackupUtils.getLogBackupDir(targetRootDir, backupId));
}
this.startTs = 0;
this.completeTs = 0;
}
public int getWorkers() {
return workers;
}
public void setWorkers(int workers) {
this.workers = workers;
}
public long getBandwidth() {
return bandwidth;
}
public void setBandwidth(long bandwidth) {
this.bandwidth = bandwidth;
}
public void setBackupTableInfoMap(Map backupTableInfoMap) {
this.backupTableInfoMap = backupTableInfoMap;
}
public HashMap> getTableSetTimestampMap() {
return tableSetTimestampMap;
}
public void setTableSetTimestampMap(HashMap> tableSetTimestampMap) {
this.tableSetTimestampMap = tableSetTimestampMap;
}
public void setType(BackupType type) {
this.type = type;
}
public void setBackupRootDir(String targetRootDir) {
this.backupRootDir = targetRootDir;
}
public void setTotalBytesCopied(long totalBytesCopied) {
this.totalBytesCopied = totalBytesCopied;
}
/**
* Set progress (0-100%)
* @param p progress value
*/
public void setProgress(int p) {
this.progress = p;
}
/**
* Get current progress
*/
public int getProgress() {
return progress;
}
public String getBackupId() {
return backupId;
}
public void setBackupId(String backupId) {
this.backupId = backupId;
}
public BackupTableInfo getBackupTableInfo(TableName table) {
return this.backupTableInfoMap.get(table);
}
public String getFailedMsg() {
return failedMsg;
}
public void setFailedMsg(String failedMsg) {
this.failedMsg = failedMsg;
}
public long getStartTs() {
return startTs;
}
public void setStartTs(long startTs) {
this.startTs = startTs;
}
public long getCompleteTs() {
return completeTs;
}
public void setCompleteTs(long endTs) {
this.completeTs = endTs;
}
public long getTotalBytesCopied() {
return totalBytesCopied;
}
public BackupState getState() {
return state;
}
public void setState(BackupState flag) {
this.state = flag;
}
public BackupPhase getPhase() {
return phase;
}
public void setPhase(BackupPhase phase) {
this.phase = phase;
}
public BackupType getType() {
return type;
}
public void setSnapshotName(TableName table, String snapshotName) {
this.backupTableInfoMap.get(table).setSnapshotName(snapshotName);
}
public String getSnapshotName(TableName table) {
return this.backupTableInfoMap.get(table).getSnapshotName();
}
public List getSnapshotNames() {
List snapshotNames = new ArrayList();
for (BackupTableInfo backupStatus : this.backupTableInfoMap.values()) {
snapshotNames.add(backupStatus.getSnapshotName());
}
return snapshotNames;
}
public Set getTables() {
return this.backupTableInfoMap.keySet();
}
public List getTableNames() {
return new ArrayList(backupTableInfoMap.keySet());
}
public void addTables(TableName[] tables) {
for (TableName table : tables) {
BackupTableInfo backupStatus = new BackupTableInfo(table, this.backupRootDir, this.backupId);
this.backupTableInfoMap.put(table, backupStatus);
}
}
public void setTables(List tables) {
this.backupTableInfoMap.clear();
for (TableName table : tables) {
BackupTableInfo backupStatus = new BackupTableInfo(table, this.backupRootDir, this.backupId);
this.backupTableInfoMap.put(table, backupStatus);
}
}
public String getBackupRootDir() {
return backupRootDir;
}
public String getTableBackupDir(TableName tableName) {
return BackupUtils.getTableBackupDir(backupRootDir, backupId, tableName);
}
public void setHLogTargetDir(String hlogTagetDir) {
this.hlogTargetDir = hlogTagetDir;
}
public String getHLogTargetDir() {
return hlogTargetDir;
}
public List getIncrBackupFileList() {
return incrBackupFileList;
}
public void setIncrBackupFileList(List incrBackupFileList) {
this.incrBackupFileList = incrBackupFileList;
}
/**
* Set the new region server log timestamps after distributed log roll
* @param newTableSetTimestampMap table timestamp map
*/
public void
setIncrTimestampMap(HashMap> newTableSetTimestampMap) {
this.tableSetTimestampMap = newTableSetTimestampMap;
}
/**
* Get new region server log timestamps after distributed log roll
* @return new region server log timestamps
*/
public HashMap> getIncrTimestampMap() {
return this.tableSetTimestampMap;
}
public TableName getTableBySnapshot(String snapshotName) {
for (Entry entry : this.backupTableInfoMap.entrySet()) {
if (snapshotName.equals(entry.getValue().getSnapshotName())) {
return entry.getKey();
}
}
return null;
}
public BackupProtos.BackupInfo toProtosBackupInfo() {
BackupProtos.BackupInfo.Builder builder = BackupProtos.BackupInfo.newBuilder();
builder.setBackupId(getBackupId());
setBackupTableInfoMap(builder);
builder.setCompleteTs(getCompleteTs());
if (getFailedMsg() != null) {
builder.setFailedMessage(getFailedMsg());
}
if (getState() != null) {
builder.setBackupState(BackupProtos.BackupInfo.BackupState.valueOf(getState().name()));
}
if (getPhase() != null) {
builder.setBackupPhase(BackupProtos.BackupInfo.BackupPhase.valueOf(getPhase().name()));
}
builder.setProgress(getProgress());
builder.setStartTs(getStartTs());
builder.setBackupRootDir(getBackupRootDir());
builder.setBackupType(BackupProtos.BackupType.valueOf(getType().name()));
builder.setWorkersNumber(workers);
builder.setBandwidth(bandwidth);
return builder.build();
}
@Override
public int hashCode() {
int hash = 33 * type.hashCode() + backupId != null ? backupId.hashCode() : 0;
if (backupRootDir != null) {
hash = 33 * hash + backupRootDir.hashCode();
}
hash = 33 * hash + state.hashCode();
hash = 33 * hash + phase.hashCode();
hash = 33 * hash + (int) (startTs ^ (startTs >>> 32));
hash = 33 * hash + (int) (completeTs ^ (completeTs >>> 32));
hash = 33 * hash + (int) (totalBytesCopied ^ (totalBytesCopied >>> 32));
if (hlogTargetDir != null) {
hash = 33 * hash + hlogTargetDir.hashCode();
}
return hash;
}
@Override
public boolean equals(Object obj) {
if (obj instanceof BackupInfo) {
BackupInfo other = (BackupInfo) obj;
try {
return Bytes.equals(toByteArray(), other.toByteArray());
} catch (IOException e) {
LOG.error(e);
return false;
}
} else {
return false;
}
}
public byte[] toByteArray() throws IOException {
return toProtosBackupInfo().toByteArray();
}
private void setBackupTableInfoMap(Builder builder) {
for (Entry entry : backupTableInfoMap.entrySet()) {
builder.addBackupTableInfo(entry.getValue().toProto());
}
}
public static BackupInfo fromByteArray(byte[] data) throws IOException {
return fromProto(BackupProtos.BackupInfo.parseFrom(data));
}
public static BackupInfo fromStream(final InputStream stream) throws IOException {
return fromProto(BackupProtos.BackupInfo.parseDelimitedFrom(stream));
}
public static BackupInfo fromProto(BackupProtos.BackupInfo proto) {
BackupInfo context = new BackupInfo();
context.setBackupId(proto.getBackupId());
context.setBackupTableInfoMap(toMap(proto.getBackupTableInfoList()));
context.setCompleteTs(proto.getCompleteTs());
if (proto.hasFailedMessage()) {
context.setFailedMsg(proto.getFailedMessage());
}
if (proto.hasBackupState()) {
context.setState(BackupInfo.BackupState.valueOf(proto.getBackupState().name()));
}
context.setHLogTargetDir(BackupUtils.getLogBackupDir(proto.getBackupRootDir(),
proto.getBackupId()));
if (proto.hasBackupPhase()) {
context.setPhase(BackupPhase.valueOf(proto.getBackupPhase().name()));
}
if (proto.hasProgress()) {
context.setProgress(proto.getProgress());
}
context.setStartTs(proto.getStartTs());
context.setBackupRootDir(proto.getBackupRootDir());
context.setType(BackupType.valueOf(proto.getBackupType().name()));
context.setWorkers(proto.getWorkersNumber());
context.setBandwidth(proto.getBandwidth());
return context;
}
private static Map toMap(List list) {
HashMap map = new HashMap<>();
for (BackupProtos.BackupTableInfo tbs : list) {
map.put(ProtobufUtil.toTableName(tbs.getTableName()), BackupTableInfo.convert(tbs));
}
return map;
}
public String getShortDescription() {
StringBuilder sb = new StringBuilder();
sb.append("{");
sb.append("ID=" + backupId).append(",");
sb.append("Type=" + getType()).append(",");
sb.append("Tables=" + getTableListAsString()).append(",");
sb.append("State=" + getState()).append(",");
Date date = null;
Calendar cal = Calendar.getInstance();
cal.setTimeInMillis(getStartTs());
date = cal.getTime();
sb.append("Start time=" + date).append(",");
if (state == BackupState.FAILED) {
sb.append("Failed message=" + getFailedMsg()).append(",");
} else if (state == BackupState.RUNNING) {
sb.append("Phase=" + getPhase()).append(",");
} else if (state == BackupState.COMPLETE) {
cal = Calendar.getInstance();
cal.setTimeInMillis(getCompleteTs());
date = cal.getTime();
sb.append("End time=" + date).append(",");
}
sb.append("Progress=" + getProgress() + "%");
sb.append("}");
return sb.toString();
}
public String getStatusAndProgressAsString() {
StringBuilder sb = new StringBuilder();
sb.append("id: ").append(getBackupId()).append(" state: ").append(getState())
.append(" progress: ").append(getProgress());
return sb.toString();
}
public String getTableListAsString() {
StringBuffer sb = new StringBuffer();
sb.append("{");
sb.append(StringUtils.join(backupTableInfoMap.keySet(), ","));
sb.append("}");
return sb.toString();
}
/**
* We use only time stamps to compare objects during sort operation
*/
@Override
public int compareTo(BackupInfo o) {
Long thisTS =
Long.valueOf(this.getBackupId().substring(this.getBackupId().lastIndexOf("_") + 1));
Long otherTS = Long.valueOf(o.getBackupId().substring(o.getBackupId().lastIndexOf("_") + 1));
return thisTS.compareTo(otherTS);
}
}