Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hive.ql.parse;
import com.google.common.base.Function;
import com.google.common.collect.Iterables;
import com.google.common.primitives.Ints;
import org.antlr.runtime.tree.Tree;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.metastore.IMetaStoreClient;
import org.apache.hadoop.hive.metastore.ReplChangeManager;
import org.apache.hadoop.hive.metastore.api.Database;
import org.apache.hadoop.hive.metastore.api.FieldSchema;
import org.apache.hadoop.hive.metastore.api.NotificationEvent;
import org.apache.hadoop.hive.metastore.messaging.AddPartitionMessage;
import org.apache.hadoop.hive.metastore.messaging.AlterPartitionMessage;
import org.apache.hadoop.hive.metastore.messaging.AlterTableMessage;
import org.apache.hadoop.hive.metastore.messaging.CreateTableMessage;
import org.apache.hadoop.hive.metastore.messaging.DropPartitionMessage;
import org.apache.hadoop.hive.metastore.messaging.DropTableMessage;
import org.apache.hadoop.hive.metastore.messaging.EventUtils;
import org.apache.hadoop.hive.metastore.messaging.InsertMessage;
import org.apache.hadoop.hive.metastore.messaging.MessageDeserializer;
import org.apache.hadoop.hive.metastore.messaging.MessageFactory;
import org.apache.hadoop.hive.metastore.messaging.PartitionFiles;
import org.apache.hadoop.hive.ql.ErrorMsg;
import org.apache.hadoop.hive.ql.QueryState;
import org.apache.hadoop.hive.ql.exec.Task;
import org.apache.hadoop.hive.ql.exec.TaskFactory;
import org.apache.hadoop.hive.ql.exec.Utilities;
import org.apache.hadoop.hive.ql.hooks.ReadEntity;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.ql.metadata.Partition;
import org.apache.hadoop.hive.ql.metadata.Table;
import org.apache.hadoop.hive.ql.plan.AlterDatabaseDesc;
import org.apache.hadoop.hive.ql.plan.AlterTableDesc;
import org.apache.hadoop.hive.ql.plan.CreateDatabaseDesc;
import org.apache.hadoop.hive.ql.plan.DDLWork;
import org.apache.hadoop.hive.ql.plan.DependencyCollectionWork;
import org.apache.hadoop.hive.ql.plan.DropTableDesc;
import org.apache.hadoop.hive.ql.plan.ExprNodeColumnDesc;
import org.apache.hadoop.hive.ql.plan.ExprNodeConstantDesc;
import org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc;
import org.apache.hadoop.hive.ql.plan.PlanUtils;
import org.apache.hadoop.hive.ql.plan.RenamePartitionDesc;
import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
import org.apache.hadoop.io.IOUtils;
import javax.annotation.Nullable;
import java.io.BufferedReader;
import java.io.BufferedWriter;
import java.io.DataOutputStream;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.io.InputStreamReader;
import java.io.OutputStreamWriter;
import java.io.Serializable;
import java.net.URI;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collections;
import java.util.HashMap;
import java.util.Iterator;
import java.util.LinkedHashMap;
import java.util.List;
import java.util.Map;
import static org.apache.hadoop.hive.ql.parse.HiveParser.*;
public class ReplicationSemanticAnalyzer extends BaseSemanticAnalyzer {
// Database name or pattern
private String dbNameOrPattern;
// Table name or pattern
private String tblNameOrPattern;
private Long eventFrom;
private Long eventTo;
private Integer maxEventLimit;
// Base path for REPL LOAD
private String path;
private static String testInjectDumpDir = null; // unit tests can overwrite this to affect default dump behaviour
private static final String dumpSchema = "dump_dir,last_repl_id#string,string";
public static final String DUMPMETADATA = "_dumpmetadata";
public enum DUMPTYPE {
BOOTSTRAP("BOOTSTRAP"),
INCREMENTAL("INCREMENTAL"),
EVENT_CREATE_TABLE("EVENT_CREATE_TABLE"),
EVENT_ADD_PARTITION("EVENT_ADD_PARTITION"),
EVENT_DROP_TABLE("EVENT_DROP_TABLE"),
EVENT_DROP_PARTITION("EVENT_DROP_PARTITION"),
EVENT_ALTER_TABLE("EVENT_ALTER_TABLE"),
EVENT_RENAME_TABLE("EVENT_RENAME_TABLE"),
EVENT_ALTER_PARTITION("EVENT_ALTER_PARTITION"),
EVENT_RENAME_PARTITION("EVENT_RENAME_PARTITION"),
EVENT_INSERT("EVENT_INSERT"),
EVENT_UNKNOWN("EVENT_UNKNOWN");
String type = null;
DUMPTYPE(String type) {
this.type = type;
}
@Override
public String toString(){
return type;
}
};
public class DumpMetaData {
// wrapper class for reading and writing metadata about a dump
// responsible for _dumpmetadata files
private DUMPTYPE dumpType;
private Long eventFrom = null;
private Long eventTo = null;
private String payload = null;
private boolean initialized = false;
private final Path dumpRoot;
private final Path dumpFile;
private Path cmRoot;
public DumpMetaData(Path dumpRoot) {
this.dumpRoot = dumpRoot;
dumpFile = new Path(dumpRoot, DUMPMETADATA);
}
public DumpMetaData(Path dumpRoot, DUMPTYPE lvl, Long eventFrom, Long eventTo, Path cmRoot){
this(dumpRoot);
setDump(lvl, eventFrom, eventTo, cmRoot);
}
public void setDump(DUMPTYPE lvl, Long eventFrom, Long eventTo, Path cmRoot){
this.dumpType = lvl;
this.eventFrom = eventFrom;
this.eventTo = eventTo;
this.initialized = true;
this.cmRoot = cmRoot;
}
public void loadDumpFromFile() throws SemanticException {
try {
// read from dumpfile and instantiate self
FileSystem fs = dumpFile.getFileSystem(conf);
BufferedReader br = new BufferedReader(new InputStreamReader(fs.open(dumpFile)));
String line = null;
if ( (line = br.readLine()) != null){
String[] lineContents = line.split("\t", 5);
setDump(DUMPTYPE.valueOf(lineContents[0]), Long.valueOf(lineContents[1]), Long.valueOf(lineContents[2]),
new Path(lineContents[3]));
setPayload(lineContents[4].equals(Utilities.nullStringOutput) ? null : lineContents[4]);
ReplChangeManager.setCmRoot(cmRoot);
} else {
throw new IOException("Unable to read valid values from dumpFile:"+dumpFile.toUri().toString());
}
} catch (IOException ioe){
throw new SemanticException(ioe);
}
}
public DUMPTYPE getDumpType() throws SemanticException {
initializeIfNot();
return this.dumpType;
}
public String getPayload() throws SemanticException {
initializeIfNot();
return this.payload;
}
public void setPayload(String payload) {
this.payload = payload;
}
public Long getEventFrom() throws SemanticException {
initializeIfNot();
return eventFrom;
}
public Long getEventTo() throws SemanticException {
initializeIfNot();
return eventTo;
}
public Path getCmRoot() {
return cmRoot;
}
public void setCmRoot(Path cmRoot) {
this.cmRoot = cmRoot;
}
public Path getDumpFilePath() {
return dumpFile;
}
public boolean isIncrementalDump() throws SemanticException {
initializeIfNot();
return (this.dumpType == DUMPTYPE.INCREMENTAL);
}
private void initializeIfNot() throws SemanticException {
if (!initialized){
loadDumpFromFile();
}
}
public void write() throws SemanticException {
writeOutput(Arrays.asList(dumpType.toString(), eventFrom.toString(), eventTo.toString(),
cmRoot.toString(), payload), dumpFile);
}
}
public ReplicationSemanticAnalyzer(QueryState queryState) throws SemanticException {
super(queryState);
}
@Override
public void analyzeInternal(ASTNode ast) throws SemanticException {
LOG.debug("ReplicationSemanticAanalyzer: analyzeInternal");
LOG.debug(ast.getName() + ":" + ast.getToken().getText() + "=" + ast.getText());
switch (ast.getToken().getType()) {
case TOK_REPL_DUMP: {
LOG.debug("ReplicationSemanticAnalyzer: analyzeInternal: dump");
initReplDump(ast);
analyzeReplDump(ast);
break;
}
case TOK_REPL_LOAD: {
LOG.debug("ReplicationSemanticAnalyzer: analyzeInternal: load");
initReplLoad(ast);
analyzeReplLoad(ast);
break;
}
case TOK_REPL_STATUS: {
LOG.debug("ReplicationSemanticAnalyzer: analyzeInternal: status");
initReplStatus(ast);
analyzeReplStatus(ast);
break;
}
default: {
throw new SemanticException("Unexpected root token");
}
}
}
private void initReplDump(ASTNode ast) {
int numChildren = ast.getChildCount();
dbNameOrPattern = PlanUtils.stripQuotes(ast.getChild(0).getText());
// skip the first node, which is always required
int currNode = 1;
while (currNode < numChildren) {
if (ast.getChild(currNode).getType() != TOK_FROM) {
// optional tblName was specified.
tblNameOrPattern = PlanUtils.stripQuotes(ast.getChild(currNode).getText());
} else {
// TOK_FROM subtree
Tree fromNode = ast.getChild(currNode);
eventFrom = Long.parseLong(PlanUtils.stripQuotes(fromNode.getChild(0).getText()));
// skip the first, which is always required
int numChild = 1;
while (numChild < fromNode.getChildCount()) {
if (fromNode.getChild(numChild).getType() == TOK_TO) {
eventTo =
Long.parseLong(PlanUtils.stripQuotes(fromNode.getChild(numChild + 1).getText()));
// skip the next child, since we already took care of it
numChild++;
} else if (fromNode.getChild(numChild).getType() == TOK_LIMIT) {
maxEventLimit =
Integer.parseInt(PlanUtils.stripQuotes(fromNode.getChild(numChild + 1).getText()));
// skip the next child, since we already took care of it
numChild++;
}
// move to the next child in FROM tree
numChild++;
}
// FROM node is always the last
break;
}
// move to the next root node
currNode++;
}
}
// REPL DUMP
private void analyzeReplDump(ASTNode ast) throws SemanticException {
LOG.debug("ReplicationSemanticAnalyzer.analyzeReplDump: " + String.valueOf(dbNameOrPattern)
+ "." + String.valueOf(tblNameOrPattern) + " from " + String.valueOf(eventFrom) + " to "
+ String.valueOf(eventTo) + " maxEventLimit " + String.valueOf(maxEventLimit));
String replRoot = conf.getVar(HiveConf.ConfVars.REPLDIR);
Path dumpRoot = new Path(replRoot, getNextDumpDir());
DumpMetaData dmd = new DumpMetaData(dumpRoot);
Path cmRoot = new Path(conf.getVar(HiveConf.ConfVars.REPLCMDIR));
Long lastReplId;
try {
if (eventFrom == null){
// bootstrap case
Long bootDumpBeginReplId = db.getMSC().getCurrentNotificationEventId().getEventId();
for (String dbName : matchesDb(dbNameOrPattern)) {
LOG.debug("ReplicationSemanticAnalyzer: analyzeReplDump dumping db: " + dbName);
Path dbRoot = dumpDbMetadata(dbName, dumpRoot);
for (String tblName : matchesTbl(dbName, tblNameOrPattern)) {
LOG.debug("ReplicationSemanticAnalyzer: analyzeReplDump dumping table: " + tblName
+ " to db root " + dbRoot.toUri());
dumpTbl(ast, dbName, tblName, dbRoot);
}
}
Long bootDumpEndReplId = db.getMSC().getCurrentNotificationEventId().getEventId();
LOG.info("Bootstrap object dump phase took from {} to {}", bootDumpBeginReplId, bootDumpEndReplId);
// Now that bootstrap has dumped all objects related, we have to account for the changes
// that occurred while bootstrap was happening - i.e. we have to look through all events
// during the bootstrap period and consolidate them with our dump.
IMetaStoreClient.NotificationFilter evFilter =
EventUtils.getDbTblNotificationFilter(dbNameOrPattern, tblNameOrPattern);
EventUtils.MSClientNotificationFetcher evFetcher =
new EventUtils.MSClientNotificationFetcher(db.getMSC());
EventUtils.NotificationEventIterator evIter = new EventUtils.NotificationEventIterator(
evFetcher, bootDumpBeginReplId,
Ints.checkedCast(bootDumpEndReplId - bootDumpBeginReplId) + 1,
evFilter );
// Now we consolidate all the events that happenned during the objdump into the objdump
while (evIter.hasNext()){
NotificationEvent ev = evIter.next();
Path evRoot = new Path(dumpRoot, String.valueOf(ev.getEventId()));
// FIXME : implement consolidateEvent(..) similar to dumpEvent(ev,evRoot)
}
LOG.info(
"Consolidation done, preparing to return {},{}->{}",
dumpRoot.toUri(), bootDumpBeginReplId, bootDumpEndReplId);
dmd.setDump(DUMPTYPE.BOOTSTRAP, bootDumpBeginReplId, bootDumpEndReplId, cmRoot);
dmd.write();
// Set the correct last repl id to return to the user
lastReplId = bootDumpEndReplId;
} else {
// get list of events matching dbPattern & tblPattern
// go through each event, and dump out each event to a event-level dump dir inside dumproot
if (eventTo == null){
eventTo = db.getMSC().getCurrentNotificationEventId().getEventId();
LOG.debug("eventTo not specified, using current event id : {}", eventTo);
} else if (eventTo < eventFrom) {
throw new Exception("Invalid event ID input received in TO clause");
}
Integer maxRange = Ints.checkedCast(eventTo - eventFrom + 1);
if ((maxEventLimit == null) || (maxEventLimit > maxRange)){
maxEventLimit = maxRange;
}
// TODO : instead of simply restricting by message format, we should eventually
// move to a jdbc-driver-stype registering of message format, and picking message
// factory per event to decode. For now, however, since all messages have the
// same factory, restricting by message format is effectively a guard against
// older leftover data that would cause us problems.
IMetaStoreClient.NotificationFilter evFilter = EventUtils.andFilter(
EventUtils.getDbTblNotificationFilter(dbNameOrPattern, tblNameOrPattern),
EventUtils.getEventBoundaryFilter(eventFrom, eventTo),
EventUtils.restrictByMessageFormat(MessageFactory.getInstance().getMessageFormat()));
EventUtils.MSClientNotificationFetcher evFetcher
= new EventUtils.MSClientNotificationFetcher(db.getMSC());
EventUtils.NotificationEventIterator evIter = new EventUtils.NotificationEventIterator(
evFetcher, eventFrom, maxEventLimit, evFilter);
lastReplId = eventTo;
while (evIter.hasNext()){
NotificationEvent ev = evIter.next();
lastReplId = ev.getEventId();
Path evRoot = new Path(dumpRoot, String.valueOf(lastReplId));
dumpEvent(ev, evRoot, cmRoot);
}
LOG.info("Done dumping events, preparing to return {},{}", dumpRoot.toUri(), lastReplId);
writeOutput(
Arrays.asList("incremental", String.valueOf(eventFrom), String.valueOf(lastReplId)),
dmd.getDumpFilePath());
dmd.setDump(DUMPTYPE.INCREMENTAL, eventFrom, lastReplId, cmRoot);
dmd.write();
}
prepareReturnValues(Arrays.asList(dumpRoot.toUri().toString(), String.valueOf(lastReplId)), dumpSchema);
setFetchTask(createFetchTask(dumpSchema));
} catch (Exception e) {
// TODO : simple wrap & rethrow for now, clean up with error codes
LOG.warn("Error during analyzeReplDump", e);
throw new SemanticException(e);
}
}
private void dumpEvent(NotificationEvent ev, Path evRoot, Path cmRoot) throws Exception {
long evid = ev.getEventId();
String evidStr = String.valueOf(evid);
ReplicationSpec replicationSpec = getNewEventOnlyReplicationSpec(evidStr);
MessageDeserializer md = MessageFactory.getInstance().getDeserializer();
switch (ev.getEventType()){
case MessageFactory.CREATE_TABLE_EVENT : {
CreateTableMessage ctm = md.getCreateTableMessage(ev.getMessage());
LOG.info("Processing#{} CREATE_TABLE message : {}", ev.getEventId(), ev.getMessage());
org.apache.hadoop.hive.metastore.api.Table tobj = ctm.getTableObj();
if (tobj == null){
LOG.debug("Event#{} was a CREATE_TABLE_EVENT with no table listed");
break;
}
Table qlMdTable = new Table(tobj);
if (qlMdTable.isView()) {
replicationSpec.setIsMetadataOnly(true);
}
Path metaDataPath = new Path(evRoot, EximUtil.METADATA_NAME);
EximUtil.createExportDump(
metaDataPath.getFileSystem(conf),
metaDataPath,
qlMdTable,
null,
replicationSpec);
Path dataPath = new Path(evRoot, "data");
Iterable files = ctm.getFiles();
if (files != null) {
// encoded filename/checksum of files, write into _files
FileSystem fs = dataPath.getFileSystem(conf);
Path filesPath = new Path(dataPath, EximUtil.FILES_NAME);
BufferedWriter fileListWriter = new BufferedWriter(
new OutputStreamWriter(fs.create(filesPath)));
try {
for (String file : files) {
fileListWriter.write(file + "\n");
}
} finally {
fileListWriter.close();
}
}
(new DumpMetaData(evRoot, DUMPTYPE.EVENT_CREATE_TABLE, evid, evid, cmRoot)).write();
break;
}
case MessageFactory.ADD_PARTITION_EVENT : {
AddPartitionMessage apm = md.getAddPartitionMessage(ev.getMessage());
LOG.info("Processing#{} ADD_PARTITION message : {}", ev.getEventId(), ev.getMessage());
Iterable ptns = apm.getPartitionObjs();
if ((ptns == null) || (!ptns.iterator().hasNext())) {
LOG.debug("Event#{} was an ADD_PTN_EVENT with no partitions");
break;
}
org.apache.hadoop.hive.metastore.api.Table tobj = apm.getTableObj();
if (tobj == null){
LOG.debug("Event#{} was a ADD_PTN_EVENT with no table listed");
break;
}
final Table qlMdTable = new Table(tobj);
Iterable qlPtns = Iterables.transform(
ptns,
new Function() {
@Nullable
@Override
public Partition apply(@Nullable org.apache.hadoop.hive.metastore.api.Partition input) {
if (input == null) {
return null;
}
try {
return new Partition(qlMdTable, input);
} catch (HiveException e) {
throw new IllegalArgumentException(e);
}
}
}
);
Path metaDataPath = new Path(evRoot, EximUtil.METADATA_NAME);
EximUtil.createExportDump(
metaDataPath.getFileSystem(conf),
metaDataPath,
qlMdTable,
qlPtns,
replicationSpec);
Iterator partitionFilesIter = apm.getPartitionFilesIter().iterator();
for (Partition qlPtn : qlPtns){
PartitionFiles partitionFiles = partitionFilesIter.next();
Iterable files = partitionFiles.getFiles();
if (files != null) {
// encoded filename/checksum of files, write into _files
Path ptnDataPath = new Path(evRoot, qlPtn.getName());
FileSystem fs = ptnDataPath.getFileSystem(conf);
Path filesPath = new Path(ptnDataPath, EximUtil.FILES_NAME);
BufferedWriter fileListWriter = new BufferedWriter(
new OutputStreamWriter(fs.create(filesPath)));
try {
for (String file : files) {
fileListWriter.write(file + "\n");
}
} finally {
fileListWriter.close();
}
}
}
(new DumpMetaData(evRoot, DUMPTYPE.EVENT_ADD_PARTITION, evid, evid, cmRoot)).write();
break;
}
case MessageFactory.DROP_TABLE_EVENT : {
LOG.info("Processing#{} DROP_TABLE message : {}", ev.getEventId(), ev.getMessage());
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_DROP_TABLE, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
break;
}
case MessageFactory.DROP_PARTITION_EVENT : {
LOG.info("Processing#{} DROP_PARTITION message : {}", ev.getEventId(), ev.getMessage());
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_DROP_PARTITION, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
break;
}
case MessageFactory.ALTER_TABLE_EVENT : {
LOG.info("Processing#{} ALTER_TABLE message : {}", ev.getEventId(), ev.getMessage());
AlterTableMessage atm = md.getAlterTableMessage(ev.getMessage());
org.apache.hadoop.hive.metastore.api.Table tobjBefore = atm.getTableObjBefore();
org.apache.hadoop.hive.metastore.api.Table tobjAfter = atm.getTableObjAfter();
if (tobjBefore.getDbName().equals(tobjAfter.getDbName()) &&
tobjBefore.getTableName().equals(tobjAfter.getTableName())){
// regular alter scenario
replicationSpec.setIsMetadataOnly(true);
Table qlMdTableAfter = new Table(tobjAfter);
Path metaDataPath = new Path(evRoot, EximUtil.METADATA_NAME);
EximUtil.createExportDump(
metaDataPath.getFileSystem(conf),
metaDataPath,
qlMdTableAfter,
null,
replicationSpec);
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_ALTER_TABLE, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
} else {
// rename scenario
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_RENAME_TABLE, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
}
break;
}
case MessageFactory.ALTER_PARTITION_EVENT : {
LOG.info("Processing#{} ALTER_PARTITION message : {}", ev.getEventId(), ev.getMessage());
AlterPartitionMessage apm = md.getAlterPartitionMessage(ev.getMessage());
org.apache.hadoop.hive.metastore.api.Table tblObj = apm.getTableObj();
org.apache.hadoop.hive.metastore.api.Partition pobjBefore = apm.getPtnObjBefore();
org.apache.hadoop.hive.metastore.api.Partition pobjAfter = apm.getPtnObjAfter();
boolean renameScenario = false;
Iterator beforeValIter = pobjBefore.getValuesIterator();
Iterator afterValIter = pobjAfter.getValuesIterator();
for ( ; beforeValIter.hasNext() ; ){
if (!beforeValIter.next().equals(afterValIter.next())){
renameScenario = true;
break;
}
}
if (!renameScenario){
// regular partition alter
replicationSpec.setIsMetadataOnly(true);
Table qlMdTable = new Table(tblObj);
List qlPtns = new ArrayList();
qlPtns.add(new Partition(qlMdTable, pobjAfter));
Path metaDataPath = new Path(evRoot, EximUtil.METADATA_NAME);
EximUtil.createExportDump(
metaDataPath.getFileSystem(conf),
metaDataPath,
qlMdTable,
qlPtns,
replicationSpec);
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_ALTER_PARTITION, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
break;
} else {
// rename scenario
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_RENAME_PARTITION, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
break;
}
}
case MessageFactory.INSERT_EVENT: {
InsertMessage insertMsg = md.getInsertMessage(ev.getMessage());
String dbName = insertMsg.getDB();
String tblName = insertMsg.getTable();
org.apache.hadoop.hive.metastore.api.Table tobj = db.getMSC().getTable(dbName, tblName);
Table qlMdTable = new Table(tobj);
Map partSpec = insertMsg.getPartitionKeyValues();
List qlPtns = null;
if (qlMdTable.isPartitioned() && !partSpec.isEmpty()) {
qlPtns = Arrays.asList(db.getPartition(qlMdTable, partSpec, false));
}
Path metaDataPath = new Path(evRoot, EximUtil.METADATA_NAME);
replicationSpec.setIsInsert(true); // Mark the replication type as insert into to avoid overwrite while import
EximUtil.createExportDump(metaDataPath.getFileSystem(conf), metaDataPath, qlMdTable, qlPtns,
replicationSpec);
Iterable files = insertMsg.getFiles();
if (files != null) {
// encoded filename/checksum of files, write into _files
Path dataPath = new Path(evRoot, EximUtil.DATA_PATH_NAME);
Path filesPath = new Path(dataPath, EximUtil.FILES_NAME);
FileSystem fs = dataPath.getFileSystem(conf);
BufferedWriter fileListWriter =
new BufferedWriter(new OutputStreamWriter(fs.create(filesPath)));
try {
for (String file : files) {
fileListWriter.write(file + "\n");
}
} finally {
fileListWriter.close();
}
}
LOG.info("Processing#{} INSERT message : {}", ev.getEventId(), ev.getMessage());
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_INSERT, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
break;
}
// TODO : handle other event types
default:
LOG.info("Dummy processing#{} message : {}", ev.getEventId(), ev.getMessage());
DumpMetaData dmd = new DumpMetaData(evRoot, DUMPTYPE.EVENT_UNKNOWN, evid, evid, cmRoot);
dmd.setPayload(ev.getMessage());
dmd.write();
break;
}
}
public static void injectNextDumpDirForTest(String dumpdir){
testInjectDumpDir = dumpdir;
}
String getNextDumpDir() {
if (conf.getBoolVar(HiveConf.ConfVars.HIVE_IN_TEST)) {
// make it easy to write .q unit tests, instead of unique id generation.
// however, this does mean that in writing tests, we have to be aware that
// repl dump will clash with prior dumps, and thus have to clean up properly.
if (testInjectDumpDir == null){
return "next";
} else {
return testInjectDumpDir;
}
} else {
return String.valueOf(System.currentTimeMillis());
// TODO: time good enough for now - we'll likely improve this.
// We may also work in something the equivalent of pid, thrid and move to nanos to ensure
// uniqueness.
}
}
/**
*
* @param dbName
* @param dumpRoot
* @return db dumped path
* @throws SemanticException
*/
private Path dumpDbMetadata(String dbName, Path dumpRoot) throws SemanticException {
Path dbRoot = new Path(dumpRoot, dbName);
try {
// TODO : instantiating FS objects are generally costly. Refactor
FileSystem fs = dbRoot.getFileSystem(conf);
Path dumpPath = new Path(dbRoot, EximUtil.METADATA_NAME);
Database dbObj = db.getDatabase(dbName);
EximUtil.createDbExportDump(fs, dumpPath, dbObj, getNewReplicationSpec());
} catch (Exception e) {
// TODO : simple wrap & rethrow for now, clean up with error codes
throw new SemanticException(e);
}
return dbRoot;
}
/**
*
* @param ast
* @param dbName
* @param tblName
* @param dbRoot
* @return tbl dumped path
* @throws SemanticException
*/
private Path dumpTbl(ASTNode ast, String dbName, String tblName, Path dbRoot) throws SemanticException {
Path tableRoot = new Path(dbRoot, tblName);
try {
URI toURI = EximUtil.getValidatedURI(conf, tableRoot.toUri().toString());
TableSpec ts = new TableSpec(db, conf, dbName + "." + tblName, null);
ExportSemanticAnalyzer.prepareExport(ast, toURI, ts, getNewReplicationSpec(), db, conf, ctx,
rootTasks, inputs, outputs, LOG);
} catch (HiveException e) {
// TODO : simple wrap & rethrow for now, clean up with error codes
throw new SemanticException(e);
}
return tableRoot;
}
// REPL LOAD
private void initReplLoad(ASTNode ast) {
int numChildren = ast.getChildCount();
path = PlanUtils.stripQuotes(ast.getChild(0).getText());
if (numChildren > 1) {
dbNameOrPattern = PlanUtils.stripQuotes(ast.getChild(1).getText());
}
if (numChildren > 2) {
tblNameOrPattern = PlanUtils.stripQuotes(ast.getChild(2).getText());
}
}
/*
* Example dump dirs we need to be able to handle :
*
* for: hive.repl.rootdir = staging/
* Then, repl dumps will be created in staging/
*
* single-db-dump: staging/blah12345 will contain a db dir for the db specified
* blah12345/
* default/
* _metadata
* tbl1/
* _metadata
* dt=20160907/
* _files
* tbl2/
* tbl3/
* unptn_tbl/
* _metadata
* _files
*
* multi-db-dump: staging/bar12347 will contain dirs for each db covered
* staging/
* bar12347/
* default/
* ...
* sales/
* ...
*
* single table-dump: staging/baz123 will contain a table object dump inside
* staging/
* baz123/
* _metadata
* dt=20150931/
* _files
*
* incremental dump : staging/blue123 will contain dirs for each event inside.
* staging/
* blue123/
* 34/
* 35/
* 36/
*/
private void analyzeReplLoad(ASTNode ast) throws SemanticException {
LOG.debug("ReplSemanticAnalyzer.analyzeReplLoad: " + String.valueOf(dbNameOrPattern) + "."
+ String.valueOf(tblNameOrPattern) + " from " + String.valueOf(path));
// for analyze repl load, we walk through the dir structure available in the path,
// looking at each db, and then each table, and then setting up the appropriate
// import job in its place.
try {
Path loadPath = new Path(path);
final FileSystem fs = loadPath.getFileSystem(conf);
if (!fs.exists(loadPath)) {
// supposed dump path does not exist.
throw new FileNotFoundException(loadPath.toUri().toString());
}
// Now, the dumped path can be one of three things:
// a) It can be a db dump, in which case we expect a set of dirs, each with a
// db name, and with a _metadata file in each, and table dirs inside that.
// b) It can be a table dump dir, in which case we expect a _metadata dump of
// a table in question in the dir, and individual ptn dir hierarchy.
// c) A dump can be an incremental dump, which means we have several subdirs
// each of which have the evid as the dir name, and each of which correspond
// to a event-level dump. Currently, only CREATE_TABLE and ADD_PARTITION are
// handled, so all of these dumps will be at a table/ptn level.
// For incremental repl, we will have individual events which can
// be other things like roles and fns as well.
// At this point, all dump dirs should contain a _dumpmetadata file that
// tells us what is inside that dumpdir.
DumpMetaData dmd = new DumpMetaData(loadPath);
boolean evDump = false;
if (dmd.isIncrementalDump()){
LOG.debug("{} contains an incremental dump", loadPath);
evDump = true;
} else {
LOG.debug("{} contains an bootstrap dump", loadPath);
}
if ((!evDump) && (tblNameOrPattern != null) && !(tblNameOrPattern.isEmpty())) {
// not an event dump, and table name pattern specified, this has to be a tbl-level dump
rootTasks.addAll(analyzeTableLoad(dbNameOrPattern, tblNameOrPattern, path, null, null, null));
return;
}
FileStatus[] srcs = LoadSemanticAnalyzer.matchFilesOrDir(fs, loadPath);
if (srcs == null || (srcs.length == 0)) {
LOG.warn("Nothing to load at {}", loadPath.toUri().toString());
return;
}
FileStatus[] dirsInLoadPath = fs.listStatus(loadPath, EximUtil.getDirectoryFilter(fs));
if ((dirsInLoadPath == null) || (dirsInLoadPath.length == 0)) {
throw new IllegalArgumentException("No data to load in path " + loadPath.toUri().toString());
}
if (!evDump){
// not an event dump, not a table dump - thus, a db dump
if ((dbNameOrPattern != null) && (dirsInLoadPath.length > 1)) {
LOG.debug("Found multiple dirs when we expected 1:");
for (FileStatus d : dirsInLoadPath) {
LOG.debug("> " + d.getPath().toUri().toString());
}
throw new IllegalArgumentException(
"Multiple dirs in "
+ loadPath.toUri().toString()
+ " does not correspond to REPL LOAD expecting to load to a singular destination point.");
}
for (FileStatus dir : dirsInLoadPath) {
analyzeDatabaseLoad(dbNameOrPattern, fs, dir);
}
} else {
// event dump, each subdir is an individual event dump.
Arrays.sort(dirsInLoadPath); // we need to guarantee that the directory listing we got is in order of evid.
Task extends Serializable> evTaskRoot = TaskFactory.get(new DependencyCollectionWork(), conf);
Task extends Serializable> taskChainTail = evTaskRoot;
int evstage = 0;
Long lastEvid = null;
Map dbsUpdated = new ReplicationSpec.ReplStateMap();
Map tablesUpdated = new ReplicationSpec.ReplStateMap();
for (FileStatus dir : dirsInLoadPath){
LOG.debug("Loading event from {} to {}.{}", dir.getPath().toUri(), dbNameOrPattern, tblNameOrPattern);
// event loads will behave similar to table loads, with one crucial difference
// precursor order is strict, and each event must be processed after the previous one.
// The way we handle this strict order is as follows:
// First, we start with a taskChainTail which is a dummy noop task (a DependecyCollectionTask)
// at the head of our event chain. For each event we process, we tell analyzeTableLoad to
// create tasks that use the taskChainTail as a dependency. Then, we collect all those tasks
// and introduce a new barrier task(also a DependencyCollectionTask) which depends on all
// these tasks. Then, this barrier task becomes our new taskChainTail. Thus, we get a set of
// tasks as follows:
//
// --->ev1.task1-- --->ev2.task1--
// / \ / \
// evTaskRoot-->*---->ev1.task2---*--> ev1.barrierTask-->*---->ev2.task2---*->evTaskChainTail
// \ /
// --->ev1.task3--
//
// Once this entire chain is generated, we add evTaskRoot to rootTasks, so as to execute the
// entire chain
String locn = dir.getPath().toUri().toString();
DumpMetaData eventDmd = new DumpMetaData(new Path(locn));
List> evTasks = analyzeEventLoad(
dbNameOrPattern, tblNameOrPattern, locn, taskChainTail,
dbsUpdated, tablesUpdated, eventDmd);
LOG.debug("evstage#{} got {} tasks", evstage, evTasks!=null ? evTasks.size() : 0);
if ((evTasks != null) && (!evTasks.isEmpty())){
Task extends Serializable> barrierTask = TaskFactory.get(new DependencyCollectionWork(), conf);
for (Task extends Serializable> t : evTasks){
t.addDependentTask(barrierTask);
LOG.debug("Added {}:{} as a precursor of barrier task {}:{}",
t.getClass(), t.getId(), barrierTask.getClass(), barrierTask.getId());
}
LOG.debug("Updated taskChainTail from {}{} to {}{}",
taskChainTail.getClass(), taskChainTail.getId(), barrierTask.getClass(), barrierTask.getId());
taskChainTail = barrierTask;
evstage++;
lastEvid = dmd.eventTo;
}
}
// Now, we need to update repl.last.id for the various parent objects that were updated.
// This update logic will work differently based on what "level" REPL LOAD was run on.
// a) If this was a REPL LOAD at a table level, i.e. both dbNameOrPattern and
// tblNameOrPattern were specified, then the table is the only thing we should
// update the repl.last.id for.
// b) If this was a db-level REPL LOAD, then we should update the db, as well as any
// tables affected by partition level operations. (any table level ops will
// automatically be updated as the table gets updated. Note - renames will need
// careful handling.
// c) If this was a wh-level REPL LOAD, then we should update every db for which there
// were events occurring, as well as tables for which there were ptn-level ops
// happened. Again, renames must be taken care of.
//
// So, what we're going to do is have each event load update dbsUpdated and tablesUpdated
// accordingly, but ignore updates to tablesUpdated & dbsUpdated in the case of a
// table-level REPL LOAD, using only the table itself. In the case of a db-level REPL
// LOAD, we ignore dbsUpdated, but inject our own, and do not ignore tblsUpdated.
// And for wh-level, we do no special processing, and use all of dbsUpdated and
// tblsUpdated as-is.
// Additional Note - although this var says "dbNameOrPattern", on REPL LOAD side,
// we do not support a pattern It can be null or empty, in which case
// we re-use the existing name from the dump, or it can be specified,
// in which case we honour it. However, having this be a pattern is an error.
// Ditto for tblNameOrPattern.
if (evstage > 0){
if ((tblNameOrPattern != null) && (!tblNameOrPattern.isEmpty())){
// if tblNameOrPattern is specified, then dbNameOrPattern will be too, and
// thus, this is a table-level REPL LOAD - only table needs updating.
// If any of the individual events logged any other dbs as having changed,
// null them out.
dbsUpdated.clear();
tablesUpdated.clear();
tablesUpdated.put(dbNameOrPattern + "." + tblNameOrPattern, lastEvid);
} else if ((dbNameOrPattern != null) && (!dbNameOrPattern.isEmpty())){
// if dbNameOrPattern is specified and tblNameOrPattern isn't, this is a
// db-level update, and thus, the database needs updating. In addition.
dbsUpdated.clear();
dbsUpdated.put(dbNameOrPattern, lastEvid);
}
}
for (String tableName : tablesUpdated.keySet()){
// weird - AlterTableDesc requires a HashMap to update props instead of a Map.
HashMap mapProp = new HashMap();
mapProp.put(
ReplicationSpec.KEY.CURR_STATE_ID.toString(),
tablesUpdated.get(tableName).toString());
AlterTableDesc alterTblDesc = new AlterTableDesc(
AlterTableDesc.AlterTableTypes.ADDPROPS, null, false);
alterTblDesc.setProps(mapProp);
alterTblDesc.setOldName(tableName);
Task extends Serializable> updateReplIdTask = TaskFactory.get(
new DDLWork(inputs, outputs, alterTblDesc), conf);
taskChainTail.addDependentTask(updateReplIdTask);
taskChainTail = updateReplIdTask;
}
for (String dbName : dbsUpdated.keySet()){
Map mapProp = new HashMap();
mapProp.put(
ReplicationSpec.KEY.CURR_STATE_ID.toString(),
dbsUpdated.get(dbName).toString());
AlterDatabaseDesc alterDbDesc = new AlterDatabaseDesc(dbName, mapProp);
Task extends Serializable> updateReplIdTask = TaskFactory.get(
new DDLWork(inputs, outputs, alterDbDesc), conf);
taskChainTail.addDependentTask(updateReplIdTask);
taskChainTail = updateReplIdTask;
}
rootTasks.add(evTaskRoot);
}
} catch (Exception e) {
// TODO : simple wrap & rethrow for now, clean up with error codes
throw new SemanticException(e);
}
}
private List> analyzeEventLoad(
String dbName, String tblName, String locn,
Task extends Serializable> precursor,
Map dbsUpdated, Map tablesUpdated,
DumpMetaData dmd) throws SemanticException {
MessageDeserializer md = MessageFactory.getInstance().getDeserializer();
switch (dmd.getDumpType()) {
case EVENT_CREATE_TABLE: {
return analyzeTableLoad(dbName, tblName, locn, precursor, dbsUpdated, tablesUpdated);
}
case EVENT_ADD_PARTITION: {
return analyzeTableLoad(dbName, tblName, locn, precursor, dbsUpdated, tablesUpdated);
}
case EVENT_DROP_TABLE: {
DropTableMessage dropTableMessage = md.getDropTableMessage(dmd.getPayload());
String actualDbName = ((dbName == null) || dbName.isEmpty() ? dropTableMessage.getDB() : dbName);
String actualTblName = ((tblName == null) || tblName.isEmpty() ? dropTableMessage.getTable() : tblName);
DropTableDesc dropTableDesc = new DropTableDesc(
actualDbName + "." + actualTblName,
null, true, true,
getNewEventOnlyReplicationSpec(String.valueOf(dmd.getEventFrom())));
Task dropTableTask = TaskFactory.get(new DDLWork(inputs, outputs, dropTableDesc), conf);
if (precursor != null){
precursor.addDependentTask(dropTableTask);
}
List> tasks = new ArrayList>();
tasks.add(dropTableTask);
LOG.debug("Added drop tbl task : {}:{}", dropTableTask.getId(), dropTableDesc.getTableName());
dbsUpdated.put(actualDbName,dmd.getEventTo());
return tasks;
}
case EVENT_DROP_PARTITION: {
try {
DropPartitionMessage dropPartitionMessage = md.getDropPartitionMessage(dmd.getPayload());
String actualDbName = ((dbName == null) || dbName.isEmpty() ? dropPartitionMessage.getDB() : dbName);
String actualTblName = ((tblName == null) || tblName.isEmpty() ? dropPartitionMessage.getTable() : tblName);
Map> partSpecs;
partSpecs =
genPartSpecs(new Table(dropPartitionMessage.getTableObj()),
dropPartitionMessage.getPartitions());
if (partSpecs.size() > 0) {
DropTableDesc dropPtnDesc = new DropTableDesc(
actualDbName + "." + actualTblName,
partSpecs, null, true,
getNewEventOnlyReplicationSpec(String.valueOf(dmd.getEventFrom())));
Task dropPtnTask =
TaskFactory.get(new DDLWork(inputs, outputs, dropPtnDesc), conf);
if (precursor != null) {
precursor.addDependentTask(dropPtnTask);
}
List> tasks = new ArrayList>();
tasks.add(dropPtnTask);
LOG.debug("Added drop ptn task : {}:{},{}", dropPtnTask.getId(),
dropPtnDesc.getTableName(), dropPartitionMessage.getPartitions());
dbsUpdated.put(actualDbName, dmd.getEventTo());
tablesUpdated.put(actualDbName + "." + actualTblName, dmd.getEventTo());
return tasks;
} else {
throw new SemanticException(
"DROP PARTITION EVENT does not return any part descs for event message :"
+ dmd.getPayload());
}
} catch (Exception e) {
if (!(e instanceof SemanticException)){
throw new SemanticException("Error reading message members", e);
} else {
throw (SemanticException)e;
}
}
}
case EVENT_ALTER_TABLE: {
return analyzeTableLoad(dbName, tblName, locn, precursor, dbsUpdated, tablesUpdated);
}
case EVENT_RENAME_TABLE: {
AlterTableMessage renameTableMessage = md.getAlterTableMessage(dmd.getPayload());
if ((tblName != null) && (!tblName.isEmpty())){
throw new SemanticException("RENAMES of tables are not supported for table-level replication");
}
try {
String oldDbName = renameTableMessage.getTableObjBefore().getDbName();
String newDbName = renameTableMessage.getTableObjAfter().getDbName();
if ((dbName != null) && (!dbName.isEmpty())){
// If we're loading into a db, instead of into the warehouse, then the oldDbName and
// newDbName must be the same
if (!oldDbName.equalsIgnoreCase(newDbName)){
throw new SemanticException("Cannot replicate an event renaming a table across"
+ " databases into a db level load " + oldDbName +"->" + newDbName);
} else {
// both were the same, and can be replaced by the new db we're loading into.
oldDbName = dbName;
newDbName = dbName;
}
}
String oldName = oldDbName + "." + renameTableMessage.getTableObjBefore().getTableName();
String newName = newDbName + "." + renameTableMessage.getTableObjAfter().getTableName();
AlterTableDesc renameTableDesc = new AlterTableDesc(oldName, newName, false);
Task renameTableTask = TaskFactory.get(new DDLWork(inputs, outputs, renameTableDesc), conf);
if (precursor != null){
precursor.addDependentTask(renameTableTask);
}
List> tasks = new ArrayList>();
tasks.add(renameTableTask);
LOG.debug("Added rename table task : {}:{}->{}", renameTableTask.getId(), oldName, newName);
dbsUpdated.put(newDbName, dmd.getEventTo()); // oldDbName and newDbName *will* be the same if we're here
tablesUpdated.remove(oldName);
tablesUpdated.put(newName, dmd.getEventTo());
// Note : edge-case here in interaction with table-level REPL LOAD, where that nukes out tablesUpdated
// However, we explicitly don't support repl of that sort, and error out above if so. If that should
// ever change, this will need reworking.
return tasks;
} catch (Exception e) {
if (!(e instanceof SemanticException)){
throw new SemanticException("Error reading message members", e);
} else {
throw (SemanticException)e;
}
}
}
case EVENT_ALTER_PARTITION: {
return analyzeTableLoad(dbName, tblName, locn, precursor, dbsUpdated, tablesUpdated);
}
case EVENT_RENAME_PARTITION: {
AlterPartitionMessage renamePtnMessage = md.getAlterPartitionMessage(dmd.getPayload());
String actualDbName = ((dbName == null) || dbName.isEmpty() ? renamePtnMessage.getDB() : dbName);
String actualTblName = ((tblName == null) || tblName.isEmpty() ? renamePtnMessage.getTable() : tblName);
Map newPartSpec = new LinkedHashMap();
Map oldPartSpec = new LinkedHashMap();
String tableName = actualDbName + "." + actualTblName;
try {
org.apache.hadoop.hive.metastore.api.Table tblObj = renamePtnMessage.getTableObj();
org.apache.hadoop.hive.metastore.api.Partition pobjBefore = renamePtnMessage.getPtnObjBefore();
org.apache.hadoop.hive.metastore.api.Partition pobjAfter = renamePtnMessage.getPtnObjAfter();
Iterator beforeValIter = pobjBefore.getValuesIterator();
Iterator afterValIter = pobjAfter.getValuesIterator();
for (FieldSchema fs : tblObj.getPartitionKeys()){
oldPartSpec.put(fs.getName(), beforeValIter.next());
newPartSpec.put(fs.getName(), afterValIter.next());
}
} catch (Exception e) {
if (!(e instanceof SemanticException)){
throw new SemanticException("Error reading message members", e);
} else {
throw (SemanticException)e;
}
}
RenamePartitionDesc renamePtnDesc = new RenamePartitionDesc(tableName, oldPartSpec, newPartSpec);
Task renamePtnTask = TaskFactory.get(new DDLWork(inputs, outputs, renamePtnDesc), conf);
if (precursor != null){
precursor.addDependentTask(renamePtnTask);
}
List> tasks = new ArrayList>();
tasks.add(renamePtnTask);
LOG.debug("Added rename ptn task : {}:{}->{}", renamePtnTask.getId(), oldPartSpec, newPartSpec);
dbsUpdated.put(actualDbName, dmd.getEventTo());
tablesUpdated.put(tableName, dmd.getEventTo());
return tasks;
}
case EVENT_INSERT: {
md = MessageFactory.getInstance().getDeserializer();
InsertMessage insertMessage = md.getInsertMessage(dmd.getPayload());
String actualDbName = ((dbName == null) || dbName.isEmpty() ? insertMessage.getDB() : dbName);
String actualTblName = ((tblName == null) || tblName.isEmpty() ? insertMessage.getTable() : tblName);
// Piggybacking in Import logic for now
return analyzeTableLoad(actualDbName, actualTblName, locn, precursor, dbsUpdated, tablesUpdated);
}
case EVENT_UNKNOWN: {
break;
}
default: {
break;
}
}
return null;
}
private Map> genPartSpecs(Table table,
List