
org.apache.phoenix.coprocessor.tasks.IndexRebuildTask Maven / Gradle / Ivy
The newest version!
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.phoenix.coprocessor.tasks;
import com.google.common.base.Strings;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.fasterxml.jackson.databind.JsonNode;
import com.fasterxml.jackson.databind.node.ObjectNode;
import org.apache.phoenix.schema.task.SystemTaskParams;
import org.apache.phoenix.util.JacksonUtil;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.mapreduce.Cluster;
import org.apache.hadoop.mapreduce.Job;
import org.apache.phoenix.coprocessor.TaskRegionObserver;
import org.apache.phoenix.jdbc.PhoenixConnection;
import org.apache.phoenix.mapreduce.index.IndexTool;
import org.apache.phoenix.query.QueryServices;
import org.apache.phoenix.schema.PTable;
import org.apache.phoenix.schema.task.Task;
import org.apache.phoenix.util.QueryUtil;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.sql.Connection;
import java.sql.SQLException;
import java.util.Map;
/**
* Task runs periodically to rebuild indexes for System.Task entries.
*
*/
public class IndexRebuildTask extends BaseTask {
public static final String INDEX_NAME = "IndexName";
public static final String JOB_ID = "JobID";
public static final String DISABLE_BEFORE = "DisableBefore";
public static final String REBUILD_ALL = "RebuildAll";
public static final Logger LOGGER = LoggerFactory.getLogger(IndexRebuildTask.class);
@Override
public TaskRegionObserver.TaskResult run(Task.TaskRecord taskRecord) {
Connection conn = null;
try {
// We have to clone the configuration because env.getConfiguration is readonly.
Configuration conf = HBaseConfiguration.create(env.getConfiguration());
conn = QueryUtil.getConnectionOnServer(env.getConfiguration());
conf.set(QueryServices.TRANSACTIONS_ENABLED, Boolean.TRUE.toString());
String data = taskRecord.getData();
if (Strings.isNullOrEmpty(taskRecord.getData())) {
data = "{}";
}
JsonNode jsonNode = JacksonUtil.getObjectReader(JsonNode.class).readValue(data);
String indexName = getIndexName(jsonNode);
if (Strings.isNullOrEmpty(indexName)) {
String str = "Index name is not found. Index rebuild cannot continue " +
"Data : " + data;
LOGGER.warn(str);
return new TaskRegionObserver.TaskResult(TaskRegionObserver.TaskResultCode.FAIL, str);
}
boolean shouldDisable = false;
if (jsonNode.has(DISABLE_BEFORE)) {
String disableBefore = jsonNode.get(DISABLE_BEFORE).toString();
if (!Strings.isNullOrEmpty(disableBefore)) {
shouldDisable = Boolean.valueOf(disableBefore);
}
}
boolean rebuildAll = false;
if (jsonNode.has(REBUILD_ALL)) {
String rebuildAllStr = jsonNode.get(REBUILD_ALL).toString();
if (!Strings.isNullOrEmpty(rebuildAllStr)) {
rebuildAll = Boolean.valueOf(rebuildAllStr);
}
}
// Run index tool async.
boolean runForeground = false;
Map.Entry indexToolRes = IndexTool
.run(conf, taskRecord.getSchemaName(), taskRecord.getTableName(), indexName,
false, taskRecord.getTenantId(), shouldDisable, rebuildAll, runForeground);
int status = indexToolRes.getKey();
if (status != 0) {
return new TaskRegionObserver.TaskResult(TaskRegionObserver.TaskResultCode.FAIL, "Index tool returned : " + status);
}
Job job = indexToolRes.getValue();
((ObjectNode) jsonNode).put(JOB_ID, job.getJobID().toString());
Task.addTask(new SystemTaskParams.SystemTaskParamsBuilder()
.setConn(conn.unwrap(PhoenixConnection.class))
.setTaskType(taskRecord.getTaskType())
.setTenantId(taskRecord.getTenantId())
.setSchemaName(taskRecord.getSchemaName())
.setTableName(taskRecord.getTableName())
.setTaskStatus(PTable.TaskStatus.STARTED.toString())
.setData(jsonNode.toString())
.setPriority(taskRecord.getPriority())
.setStartTs(taskRecord.getTimeStamp())
.setEndTs(null)
.setAccessCheckEnabled(true)
.build());
// It will take some time to finish, so we will check the status in a separate task.
return null;
}
catch (Throwable t) {
LOGGER.warn("Exception while running index rebuild task. " +
"It will be retried in the next system task table scan : " +
taskRecord.getSchemaName() + "." + taskRecord.getTableName() +
" with tenant id " + (taskRecord.getTenantId() == null ? " IS NULL" : taskRecord.getTenantId()) +
" and data " + taskRecord.getData(), t);
return new TaskRegionObserver.TaskResult(TaskRegionObserver.TaskResultCode.FAIL, t.toString());
} finally {
if (conn != null) {
try {
conn.close();
} catch (SQLException e) {
LOGGER.debug("IndexRebuildTask can't close connection");
}
}
}
}
private String getIndexName(JsonNode jsonNode) {
String indexName = null;
// Get index name from data column.
if (jsonNode.has(INDEX_NAME)) {
indexName = jsonNode.get(INDEX_NAME).toString().replaceAll("\"", "");
}
return indexName;
}
private String getJobID(String data) throws JsonProcessingException {
if (Strings.isNullOrEmpty(data)) {
data = "{}";
}
JsonNode jsonNode = JacksonUtil.getObjectReader().readTree(data);
String jobId = null;
if (jsonNode.has(JOB_ID)) {
jobId = jsonNode.get(JOB_ID).textValue().replaceAll("\"", "");
}
return jobId;
}
@Override
public TaskRegionObserver.TaskResult checkCurrentResult(Task.TaskRecord taskRecord)
throws Exception {
String jobID = getJobID(taskRecord.getData());
if (jobID != null) {
Configuration conf = HBaseConfiguration.create(env.getConfiguration());
Configuration configuration = HBaseConfiguration.addHbaseResources(conf);
Cluster cluster = new Cluster(configuration);
Job job = cluster.getJob(org.apache.hadoop.mapreduce.JobID.forName(jobID));
if (job == null) {
return new TaskRegionObserver.TaskResult(TaskRegionObserver.TaskResultCode.SKIPPED, "");
}
if (job != null && job.isComplete()) {
if (job.isSuccessful()) {
LOGGER.warn("IndexRebuildTask checkCurrentResult job is successful "
+ taskRecord.getTableName());
return new TaskRegionObserver.TaskResult(TaskRegionObserver.TaskResultCode.SUCCESS, "");
} else {
return new TaskRegionObserver.TaskResult(TaskRegionObserver.TaskResultCode.FAIL,
"Index is DISABLED");
}
}
}
return null;
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy