All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.kylin.engine.mr.common.HadoopCmdOutput Maven / Gradle / Ivy

There is a newer version: 3.1.3
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 * 
 *     http://www.apache.org/licenses/LICENSE-2.0
 * 
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
*/

package org.apache.kylin.engine.mr.common;

import java.io.IOException;
import java.util.Collections;
import java.util.HashMap;
import java.util.Map;

import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapreduce.Counters;
import org.apache.hadoop.mapreduce.FileSystemCounter;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.JobStatus;
import org.apache.hadoop.mapreduce.TaskCompletionEvent;
import org.apache.hadoop.mapreduce.TaskCounter;
import org.apache.kylin.common.KylinConfig;
import org.apache.kylin.engine.mr.steps.FactDistinctColumnsMapper.RawDataCounter;
import org.apache.kylin.job.constant.ExecutableConstants;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

/**
 * @author xduo
 * 
 */
public class HadoopCmdOutput {

    protected static final Logger logger = LoggerFactory.getLogger(HadoopCmdOutput.class);

    private final StringBuilder output;
    private final Job job;

    public HadoopCmdOutput(Job job, StringBuilder output) {
        super();
        this.job = job;
        this.output = output;
    }

    public String getMrJobId() {
        return getInfo().get(ExecutableConstants.MR_JOB_ID);
    }

    public Map getInfo() {
        if (job != null) {
            Map status = new HashMap();
            if (null != job.getJobID()) {
                status.put(ExecutableConstants.MR_JOB_ID, job.getJobID().toString());
            }
            if (null != job.getTrackingURL()) {
                status.put(ExecutableConstants.YARN_APP_URL, job.getTrackingURL().toString());
            }
            return status;
        } else {
            return Collections.emptyMap();
        }
    }

    private String mapInputRecords;
    private String hdfsBytesWritten;
    private String rawInputBytesRead;

    public String getMapInputRecords() {
        return mapInputRecords;
    }

    public String getHdfsBytesWritten() {
        return hdfsBytesWritten;
    }

    public String getRawInputBytesRead() {
        return rawInputBytesRead;
    }

    public void updateJobCounter() {
        try {
            Counters counters = job.getCounters();
            if (counters == null) {
                String errorMsg = "no counters for job " + getMrJobId();
                logger.warn(errorMsg);
                output.append(errorMsg);
            } else {
                this.output.append(counters.toString()).append("\n");
                logger.debug(counters.toString());

                mapInputRecords = String.valueOf(counters.findCounter(TaskCounter.MAP_INPUT_RECORDS).getValue());
                rawInputBytesRead = String.valueOf(counters.findCounter(RawDataCounter.BYTES).getValue());

                String outputFolder = job.getConfiguration().get("mapreduce.output.fileoutputformat.outputdir",
                        KylinConfig.getInstanceFromEnv().getHdfsWorkingDirectory());
                logger.debug("outputFolder is " + outputFolder);
                Path outputPath = new Path(outputFolder);
                String fsScheme = outputPath.getFileSystem(job.getConfiguration()).getScheme();
                long bytesWritten = counters.findCounter(fsScheme, FileSystemCounter.BYTES_WRITTEN).getValue();
                if (bytesWritten == 0) {
                    logger.debug("Seems no counter found for " + fsScheme);
                    bytesWritten = counters.findCounter("FileSystemCounters", "HDFS_BYTES_WRITTEN").getValue();
                }
                hdfsBytesWritten = String.valueOf(bytesWritten);
            }
            JobStatus jobStatus = job.getStatus();
            if (jobStatus.getState() == JobStatus.State.FAILED) {
                logger.warn("Job Diagnostics:" + jobStatus.getFailureInfo());
                output.append("Job Diagnostics:").append(jobStatus.getFailureInfo()).append("\n");
                TaskCompletionEvent taskEvent = getOneTaskFailure(job);
                if (taskEvent != null) {
                    String[] fails = job.getTaskDiagnostics(taskEvent.getTaskAttemptId());
                    logger.warn("Failure task Diagnostics:");
                    output.append("Failure task Diagnostics:").append("\n");
                    for (String failure : fails) {
                        logger.warn(failure);
                        output.append(failure).append("\n");
                    }
                }
            }
        } catch (Exception e) {
            logger.error(e.getLocalizedMessage(), e);
            output.append(e.getLocalizedMessage());
        }
    }

    private TaskCompletionEvent getOneTaskFailure(Job job) throws IOException, InterruptedException {
        TaskCompletionEvent lastEvent = null;
        int index = 0;
        int failCount = 0;
        TaskCompletionEvent[] events = job.getTaskCompletionEvents(index);
        //This returns either nothing (if no task executions or no exceptions at all) or the last failure event within a subset of the exceptions from the first
        //index at which exceptions are found in the task completion events
        if (events == null) {
            return lastEvent;
        }
        while (events.length > 0 && failCount == 0) {
            for (TaskCompletionEvent event : events) {
                if (event.getStatus().equals(TaskCompletionEvent.Status.FAILED)) {
                    failCount++;
                    lastEvent = event;
                }
            }
            index += 10;
            events = job.getTaskCompletionEvents(index);
        }
        return lastEvent;
    }
}




© 2015 - 2025 Weber Informatics LLC | Privacy Policy